Discover how a bimodal integration strategy can address the major data management challenges facing your organization today.
Get the Report →Connect to Databricks Data in Airbyte ELT Pipelines
Use CData Connect Cloud to build ELT pipelines for Databricks data in Airbyte.
Airbyte empowers users to load your data into any data warehouse, data lake, or database. When combined with CData Connect Cloud, Airbyte users can create Extract, Load, Transform (ELT) pipelines directly from live Databricks data. This article illustrates the process of connecting to Databricks through Connect Cloud and constructing ELT pipelines for Databricks data within Airbyte.
CData Connect Cloud offers a dedicated SQL Server interface for Databricks, facilitating data querying without the need for data replication to a native database. With built-in optimized data processing capabilities, CData Connect Cloud efficiently directs all supported SQL operations, including filters and JOINs, straight to Databricks. This harnesses server-side processing to swiftly retrieve the desired Databricks data.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Configure Databricks Connectivity for Airbyte
Connectivity to Databricks from Airbyte is made possible through CData Connect Cloud. To work with Databricks data from Airbyte, we start by creating and configuring a Databricks connection.
- Log into Connect Cloud, click Connections and click Add Connection
- Select "Databricks" from the Add Connection panel
-
Enter the necessary authentication properties to connect to Databricks.
To connect to a Databricks cluster, set the properties as described below.
Note: The needed values can be found in your Databricks instance by navigating to Clusters, and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.
- Server: Set to the Server Hostname of your Databricks cluster.
- HTTPPath: Set to the HTTP Path of your Databricks cluster.
- Token: Set to your personal access token (this value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).
- Click Create & Test
- Navigate to the Permissions tab in the Add Databricks Connection page and update the User-based permissions.
Add a Personal Access Token
If you are connecting from a service, application, platform, or framework that does not support OAuth authentication, you can create a Personal Access Token (PAT) to use for authentication. Best practices would dictate that you create a separate PAT for each service, to maintain granularity of access.
- Click on your username at the top right of the Connect Cloud app and click User Profile.
- On the User Profile page, scroll down to the Personal Access Tokens section and click Create PAT.
- Give your PAT a name and click Create.
- The personal access token is only visible at creation, so be sure to copy it and store it securely for future use.
Connect to Databricks from Airbyte
To establish a connection from Airbyte to CData Connect Cloud, follow these steps.
- Log in to your Airbyte account
- On the left panel, click Sources, then Add New Source
- Set Source Type to MSSQL Server to connect the TDS endpoint
- Set Source Name
- Set Host URL to tds.cdata.com
- Set Port to 14333
- Set Database to the name of the connection you previously configured, e.g. Databricks1.
- Set Username to your Connect Cloud username
- Set SSL Method to Encrypted (trust server certificate), leave the Replication Method as standard, and set SSH Tunnel Method to No Tunnel
- (Optional) Set Schema to anything you want to apply to the source
- Set Password to your Connect Cloud PAT
- (Optional) Enter any needed JBDC URL Params
- Click Test and Save to create the data source.
Create ELT Pipelines for Databricks Data
To connect Databricks data with a new destination, click Connections and then Set Up Connection to connect to your destination. Select the source created above and your desired destination, then allow Airbyte to process. When it is done, your connection is ready for use.
Get CData Connect Cloud
To get live data access to 100+ SaaS, Big Data, and NoSQL sources directly from Airbyte, try CData Connect Cloud today!