We are proud to share our inclusion in the 2024 Gartner Magic Quadrant for Data Integration Tools. We believe this recognition reflects the differentiated business outcomes CData delivers to our customers.
Get the Report →Enable the Databricks JDBC Driver in KNIME
Use standard data access components in KNIME to create charts and reports with Databricks data.
One of the strengths of the CData JDBC Driver for Databricks is its cross-platform support, enabling integration with major BI tools. Follow the procedure below to access Databricks data in KNIME and to create a chart from Databricks data using the report designer.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Define a New JDBC Connection to Databricks Data
- Go to File -> Preferences -> KNIME -> Databases
- Click Add File and add the cdata.jdbc.databricks.jar. The driver JAR file is located in the lib subfolder of the installation directory.
- Click Find driver classes
- Set the ID and the Name of the connection (you can set any values you prefer for these fields as they are not restricted)
- Set the URL template. A typical JDBC URL is provided below:
jdbc:databricks:Server=127.0.0.1;Port=443;TransportMode=HTTP;HTTPPath=MyHTTPPath;UseSSL=True;User=MyUser;Password=MyPassword;
- Click OK to close the configuration section.
- Under Node Repository go to DB -> Connection -> drag and drop DB Connector. Double-click on it and change the driver name to the driver you just configured, in this case, Databricks (ID: Databricks). The Database URL should change automatically.
- Click Apply and OK to save changes.
- Right-click on the connector and select Execute. You will see that the connector will allow you to redirect to the browser where you will need to log in and allow access. After that, you will get connected successfully.
-
Double-click on your DB Query Reader and click the refresh button to load the metadata. Create an SQL Statement and click Evaluate. After clicking Evaluate you will be able to see the records requested. To learn more about the tables/views that are listed in our driver please refer to our Data Model.
For assistance in constructing the JDBC URL, use the connection string designer built into the Databricks data JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.
java -jar cdata.jdbc.databricks.jar

Additionally, please refer to our documentation to learn more about Connection Properties
Now you can go to the Read/Write section in Node Repository and get a DB Query Reader, to be able to execute a query.
Get Started Today
Download a free, 30-day trial of the CData JDBC Driver for Databricks and start building Databricks-connected charts and reports with KNIME. Reach out to our Support Team if you have any questions.