Extract, Transform, and Load Azure Data Lake Storage Data in Informatica PowerCenter

Ready to get started?

Download for a free trial:

Download Now

Learn more:

Azure Data Lake Storage ODBC Driver

The Azure Data Lake Storage ODBC Driver is a powerful tool that allows you to connect with live data from Azure Data Lake Storage, directly from any applications that support ODBC connectivity.

Access Azure Data Lake Storage data like you would a database - read, write, and update Azure Data Lake Storage ADLSData, etc. through a standard ODBC Driver interface.



Create a simple Workflow for Azure Data Lake Storage data in Informatica PowerCenter.

Informatica provides a powerful, elegant means of transporting and transforming your data. By utilizing the CData ODBC Driver for Azure Data Lake Storage, you are gaining access to a driver based on industry-proven standards that integrates seamlessly with Informatica's powerful data transportation and manipulation features. This tutorial shows how to create a simple Workflow in Informatica PowerCenter to extract Azure Data Lake Storage data and load it into a flat file.

Add Azure Data Lake Storage as an ODBC Data Source

If you have not already, install the driver on the PowerCenter server and client machines. On both machines, specify the connection properties in an ODBC DSN (data source name). This is the last step of the driver installation. You can use the Microsoft ODBC Data Source Administrator to create and configure ODBC DSNs.

Authenticating to a Gen 1 DataLakeStore Account

Gen 1 uses OAuth 2.0 in Azure AD for authentication.

For this, an Active Directory web application is required. You can create one as follows:

  1. Sign in to your Azure Account through the .
  2. Select "Azure Active Directory".
  3. Select "App registrations".
  4. Select "New application registration".
  5. Provide a name and URL for the application. Select Web app for the type of application you want to create.
  6. Select "Required permissions" and change the required permissions for this app. At a minimum, "Azure Data Lake" and "Windows Azure Service Management API" are required.
  7. Select "Key" and generate a new key. Add a description, a duration, and take note of the generated key. You won't be able to see it again.

To authenticate against a Gen 1 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen1.
  • Account: Set this to the name of the account.
  • OAuthClientId: Set this to the application Id of the app you created.
  • OAuthClientSecret: Set this to the key generated for the app you created.
  • TenantId: Set this to the tenant Id. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

Authenticating to a Gen 2 DataLakeStore Account

To authenticate against a Gen 2 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen2.
  • Account: Set this to the name of the account.
  • FileSystem: Set this to the file system which will be used for this account.
  • AccessKey: Set this to the access key which will be used to authenticate the calls to the API. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

Create an ETL Workflow in PowerCenter

Follow the steps below to create a workflow in PowerCenter to pull Azure Data Lake Storage data and push it into a flat file.

Create a Source Using the ODBC Driver

  1. In the powrmart.ini file found in %INFA_HOME%\clients\PowerCenterClient\client\bin, add the following line to the ODBCDLL section:
    CData=PMODBC.DLL
  2. In PowerCenter Workflow Manager, add a new ODBC relational connection, set the user name and password properties to the user for the machine on which the ODBC Driver is installed, and set the connect string to the System DSN for the driver (CData ADLS Sys).
  3. In PowerCenter Designer, connect to your repository and open your folder.
  4. Select the Source Analyzer, click the sources menu, and select Import from Database...
  5. In the drop-down menu for ODBC data source, select the DSN you previously configured (CData ADLS Sys).
  6. Click connect and select the tables and views to include.
  7. Click OK.

Create a Flat File Target Based on the Source

  1. Select the Targets Analyzer and drag and drop the previously created source onto the workspace. Using the existing source copies the columns into the target.
  2. Right-click the new target, click edit, and change the database type to flat file.

Create a Mapping to Between Azure Data Lake Storage Data and a Flat File

  1. Click on the Mapping Designer.
  2. Drag the source and target to the workspace (name the new mapping, if prompted).
  3. Right-click on the workspace and select Autolink by Name.
  4. Drag the columns from the source qualifier to the target definition.
  5. Save the folder (Ctrl + S).

Create Workflow Based on the Mapping

With the source, target, and mapping created and saved, you are now ready to create the workflow.

  1. Right-click the mapping and select generate workflow to open the Workflow Generation wizard.
  2. Create a workflow with a non-reusable session.
  3. Ensure that you have properly configured the connection object (to the relational connection we created earlier) and set the prefixes.
  4. Configure the Connection as needed.
  5. Review the Workflow and click Finish.

With a workflow created, you can open the PowerCenter Workflow Manager to access and start the workflow, quickly transferring Azure Data Lake Storage data into a flat file. With the ODBC Driver for Azure Data Lake Storage, you can configure sources and targets in PowerCenter to integrate Azure Data Lake Storage data into any of the elegant and powerful features in Informatica PowerCenter.