Hitachi Content Platform (HCP) on S3

Our HCP connection allows you to connect directly to your HCP S3 object store, allowing you to query over it using SQL or Spark, with external connection possibilities too.

Create a standard file connection

Create a lakehouse connection

Standard file connection

Step 1: Click the NDP Fabric Builder icon.

Step 2: Click the '+' next to Virtual File Sources to launch the wizard

Step 3: Name your data source e.g. HCP_FILES

Step 4: Click Create

Step 5: Click the '+' next to the newly created Virtual File Source container

Step 6: Select 'File Store' and click Next

Step 7: Complete the form

  1. File source: Choose your File Source Type
  2. Storage: Choose Amazon S3 
  3. Table name: Provide any name (without spaces - use underscore)
  4. File path: Provide your s3 path (eg. s3a://zetaris-lightning-test/csv-data/tpc-h/customer.csv)
  5. Endpoint: <accountname>.hitachivantara.com
  6. Access key: <awsAccessKeyId>
  7. Security key: <awsSecretAccessKey>
  8. Delimiter: (only or CSV file types)
  9. Parameters
    1. header - true
    2. inferSchema - true (Optional)

Step 8: Click Next

Step 9: Click Create

Lakehouse connection

Step 1: Click the NDP Fabric Builder icon.

Step 2: Click the '+' next to Data Lake Warehouse Source to launch the wizard.

Step 3: Complete the necessary fields as shown below

Name: <AnyName> (all one word)

Storage: Select AWS S3

Storage Path: s3a://path/to/lakehouse

NOTE: For Storage Path, a directory/folder must be created inside the HCP bucket, and called in the string

awsAccessKeyId: <awsAccessKeyId>

awsSecretAccessKey: <awsSecretAccessKey>

s3Endpoint: <accountname>.hitachivantara.com

Step 4: Click Create