Integrate Azure Data Lake Storage Data in Your Informatica Cloud Instance

詳細情報をご希望ですか?

無償トライアル:

ダウンロードへ

製品の詳細情報へ:

Azure Data Lake Storage JDBC Driver

Azure Data Lake Storage データに連携するJava アプリケーションを素早く、簡単に開発できる便利なドライバー。



Use the CData JDBC Driver for Azure Data Lake Storage with the Informatica Cloud Secure Agent to access live Azure Data Lake Storage data from Informatica Cloud.

Informatica Cloud allows you to perform extract, transform, and load (ETL) tasks in the cloud. With the Cloud Secure Agent and the CData JDBC Driver for Azure Data Lake Storage, you get live access to Azure Data Lake Storage data, directly within Informatica Cloud. In this article, we will walk through downloading and registering the Cloud Secure Agent, connecting to Azure Data Lake Storage through the JDBC Driver and generating a mapping that can be used in any Informatica Cloud process.

Informatica Cloud Secure Agent

To work with the Azure Data Lake Storage data through the JDBC Driver, install the Cloud Secure Agent.

  1. Navigate to the Administrator page in Informatica Cloud
  2. Select the Runtime Environments tab
  3. Click "Download Secure Agent"
  4. Make note of the Install Token
  5. Run the installer on the client machine and register the Cloud Secure Agent with your username and install token

NOTE: It may take some time for all of the Cloud Secure Agent services to get up and running.

Connecting to the Azure Data Lake Storage JDBC Driver

With the Cloud Secure Agent installed and running, you are ready to connect to Azure Data Lake Storage through the JDBC Driver. Start by clicking the Connections tab and clicking New Connection. Fill in the following properties for the connection:

  • Connection Name: Name your connection (i.e.: CData Azure Data Lake Storage Connection)
  • Type: Select "JDBC_IC (Informatica Cloud)"
  • Runtime Environment: Select the runtime environment where you installed the Cloud Secure Agent
  • JDBC Connection URL: Set this to the JDBC URL for Azure Data Lake Storage. Your URL will look similar to the following:

    jdbc:adls:Schema=ADLSGen2;Account=myAccount;FileSystem=myFileSystem;AccessKey=myAccessKey;InitiateOAuth=GETANDREFRESH;

    Authenticating to a Gen 1 DataLakeStore Account

    Gen 1 uses OAuth 2.0 in Azure AD for authentication.

    For this, an Active Directory web application is required. You can create one as follows:

    1. Sign in to your Azure Account through the .
    2. Select "Azure Active Directory".
    3. Select "App registrations".
    4. Select "New application registration".
    5. Provide a name and URL for the application. Select Web app for the type of application you want to create.
    6. Select "Required permissions" and change the required permissions for this app. At a minimum, "Azure Data Lake" and "Windows Azure Service Management API" are required.
    7. Select "Key" and generate a new key. Add a description, a duration, and take note of the generated key. You won't be able to see it again.

    To authenticate against a Gen 1 DataLakeStore account, the following properties are required:

    • Schema: Set this to ADLSGen1.
    • Account: Set this to the name of the account.
    • OAuthClientId: Set this to the application Id of the app you created.
    • OAuthClientSecret: Set this to the key generated for the app you created.
    • TenantId: Set this to the tenant Id. See the property for more information on how to acquire this.
    • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

    Authenticating to a Gen 2 DataLakeStore Account

    To authenticate against a Gen 2 DataLakeStore account, the following properties are required:

    • Schema: Set this to ADLSGen2.
    • Account: Set this to the name of the account.
    • FileSystem: Set this to the file system which will be used for this account.
    • AccessKey: Set this to the access key which will be used to authenticate the calls to the API. See the property for more information on how to acquire this.
    • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

    Built-In Connection String Designer

    For assistance in constructing the JDBC URL, use the connection string designer built into the Azure Data Lake Storage JDBC Driver. Either double-click the .jar file or execute the .jar file from the command-line.

    java -jar cdata.jdbc.adls.jar

    Fill in the connection properties and copy the connection string to the clipboard.

  • JDBC Jar Directory: Set this to the lib folder in the installation location for the JDBC Driver (on Windows, typically C:\Program Files\CData\CData JDBC Driver for Azure Data Lake Storage\)
  • Driver Class: Set this to cdata.jdbc.adls.ADLSDriver
  • Username: Set this to a placeholder value (since Azure Data Lake Storage does not require a username)
  • Password: Set this to a placeholder value (since Azure Data Lake Storage does not require a password)

Create a Mapping for Azure Data Lake Storage Data

With the connection to Azure Data Lake Storage configured, we can now access Azure Data Lake Storage data in any Informatica process. The steps below walk through creating a mapping for Azure Data Lake Storage to another data target.

  1. Navigate to the Data Integration page
  2. Click New.. and select Mapping from the Mappings tab
  3. Click the Source Object and in the Source tab, select the Connection and set the Source Type
  4. Click "Select" to choose the table to map
  5. In the Fields tab, select the fields from the Azure Data Lake Storage table to map
  6. Click the Target object and configure the Target source, table and fields. In the Field Mapping tab, map the source fields to the target fields.

With the mapping configured, you are ready to start integrating live Azure Data Lake Storage data with any of the supported connections in Informatica Cloud. Download a free, 30-day trial of the CData JDBC Driver for Azure Data Lake Storage and start working with your live Azure Data Lake Storage data in Informatica Cloud today.