Build Pipelines with Live Azure Data Lake Storage Data in Google Cloud Data Fusion (via CData Connect Cloud)



Use CData Connect Cloud to connect to Azure Data Lake Storage from Google Cloud Data Fusion, enabling the integration of live Azure Data Lake Storage data into the building and management of effective data pipelines.

Google Cloud Data Fusion simplifies building and managing data pipelines by offering a visual interface to connect, transform, and move data across various sources and destinations, streamlining data integration processes. When combined with CData Connect Cloud, it provides access to Azure Data Lake Storage data for building and managing ELT/ETL data pipelines. This article explains how to use CData Connect Cloud to create a live connection to Azure Data Lake Storage and how to connect and access live Azure Data Lake Storage data from the Cloud Data Fusion platform.

Configure Azure Data Lake Storage Connectivity for Cloud Data Fusion

Connectivity to Azure Data Lake Storage from Cloud Data Fusion is made possible through CData Connect Cloud. To work with Azure Data Lake Storage data from Cloud Data Fusion, we start by creating and configuring a Azure Data Lake Storage connection.

  1. Log into Connect Cloud, click Sources, and then click Add Connection
  2. Select "Azure Data Lake Storage" from the Add Connection panel
  3. Enter the necessary authentication properties to connect to Azure Data Lake Storage.

    Authenticating to a Gen 1 DataLakeStore Account

    Gen 1 uses OAuth 2.0 in Entra ID (formerly Azure AD) for authentication.

    For this, an Active Directory web application is required. You can create one as follows:

    1. Sign in to your Azure Account through the .
    2. Select "Entra ID" (formerly Azure AD).
    3. Select "App registrations".
    4. Select "New application registration".
    5. Provide a name and URL for the application. Select Web app for the type of application you want to create.
    6. Select "Required permissions" and change the required permissions for this app. At a minimum, "Azure Data Lake" and "Windows Azure Service Management API" are required.
    7. Select "Key" and generate a new key. Add a description, a duration, and take note of the generated key. You won't be able to see it again.

    To authenticate against a Gen 1 DataLakeStore account, the following properties are required:

    • Schema: Set this to ADLSGen1.
    • Account: Set this to the name of the account.
    • OAuthClientId: Set this to the application Id of the app you created.
    • OAuthClientSecret: Set this to the key generated for the app you created.
    • TenantId: Set this to the tenant Id. See the property for more information on how to acquire this.
    • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

    Authenticating to a Gen 2 DataLakeStore Account

    To authenticate against a Gen 2 DataLakeStore account, the following properties are required:

    • Schema: Set this to ADLSGen2.
    • Account: Set this to the name of the account.
    • FileSystem: Set this to the file system which will be used for this account.
    • AccessKey: Set this to the access key which will be used to authenticate the calls to the API. See the property for more information on how to acquire this.
    • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.
  4. Click Create & Test
  5. Navigate to the Permissions tab in the Add Azure Data Lake Storage Connection page and update the User-based permissions.

Add a Personal Access Token

When connecting to Connect Cloud through the REST API, the OData API, or the Virtual SQL Server, a Personal Access Token (PAT) is used to authenticate the connection to Connect Cloud. It is best practice to create a separate PAT for each service to maintain granularity of access.

  1. Click on the Gear icon () at the top right of the Connect Cloud app to open the settings page.
  2. On the Settings page, go to the Access Tokens section and click Create PAT.
  3. Give the PAT a name and click Create.
  4. The personal access token is only visible at creation, so be sure to copy it and store it securely for future use.

With the connection configured and a PAT generated, you are ready to connect to Azure Data Lake Storage data from Cloud Data Fusion.

Connecting to Azure Data Lake Storage from Cloud Data Fusion

Follow these steps to establish a connection from Cloud Data Fusion to Azure Data Lake Storage through the CData Connect Cloud JDBC driver:

  1. Download and install the CData Connect Cloud JDBC driver:
    1. Open the Integrations page of CData Connect Cloud.
    2. Search for and select JDBC.
    3. Download and run the setup file.
    4. When the installation is complete, copy the JAR file(cdata.jdbc.connect.jar) from the installation directory (e.g., C:\Program Files\CData\JDBC Driver for CData Connect\lib).
  2. Log into Cloud Data Fusion.
  3. Click the green "+" button at the top right to add an entity.
  4. Under Driver, click Upload.
  5. Now, upload the CData Connect Cloud JDBC driver (JAR file).
  6. Enter the driver settings:
    • Name: Enter the name of the driver
    • Class name: Enter "cdata.jdbc.connect.ConnectDriver"
    • Version: Enter the driver version
    • Description (optional): Enter a description for the driver
  7. Click on Finish.
  8. Enter source configuration settings:
    • Label: Helps to identify the connection
    • JDBC driver name: Enter the JDBC driver name to identify the driver configured in Step 6.
    • Connection string: Enter the JDBC connection string, for example:
      jdbc:connect:AuthScheme=Basic;user=username;password=PAT;
    • User: Enter your CData Connect Cloud username, displayed in the top-right corner of the CData Connect Cloud interface. For example, "[email protected]"
    • Password: Enter the PAT you generated on the Settings page.
  9. Click Validate in the top right corner.
  10. If the connection is successful, you can manage the pipeline by editing it through the UI.
  11. Run the pipepline created.

Troubleshooting

Please be aware that there is a known issue in Cloud Data Fusion where "int" types from source data are automatically cast as "long".

Live Access to Azure Data Lake Storage Data from Cloud Applications

Now you have a direct connection to live Azure Data Lake Storage data from from Google Cloud Data Fusion. You can create more connections to ensure a smooth movement of data across various sources and destinations, thereby streamlining data integration processes - all without replicating Azure Data Lake Storage data.

To get real-time data access to 100+ SaaS, Big Data, and NoSQL sources (including Azure Data Lake Storage) directly from your cloud applications, explore the CData Connect Cloud.

Ready to get started?

Learn more about CData Connect Cloud or sign up for free trial access:

Free Trial