We are proud to share our inclusion in the 2024 Gartner Magic Quadrant for Data Integration Tools. We believe this recognition reflects the differentiated business outcomes CData delivers to our customers.
Get the Report →Edit and Search External Databricks Objects in Salesforce Connect
Use CData Connect Cloud to securely provide OData feeds of Databricks data to smart devices and cloud-based applications. Use the CData Connect and Salesforce Connect to create Databricks Data objects that you can access from apps and the dashboard.
CData Connect Cloud, enables you to access Databricks data from cloud-based applications like the Salesforce console and mobile applications like the Salesforce Mobile App. In this article, you will use CData Connect Cloud and Salesforce Connect to access external Databricks objects alongside standard Salesforce objects.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Connect to Databricks from Salesforce
To work with live Databricks data in Salesforce Connect, we need to connect to Databricks from Connect Cloud, provide user access to the connection, and create OData endpoints for the Databricks data.
(Optional) Add a New Connect Cloud User
As needed, create Users to connect to Databricks through Connect Cloud.
- Navigate to the Users page and click Invite Users
- Enter the new user's email address and click Send to invite the user
- You can review and edit users from the Users page
Add a Personal Access Token
If you are connecting from a service, application, platform, or framework that does not support OAuth authentication, you can create a Personal Access Token (PAT) to use for authentication. Best practices would dictate that you create a separate PAT for each service, to maintain granularity of access.
- Click on your username at the top right of the Connect Cloud app and click User Profile.
- On the User Profile page, scroll down to the Personal Access Tokens section and click Create PAT.
- Give your PAT a name and click Create.
- The personal access token is only visible at creation, so be sure to copy it and store it securely for future use.
Connect to Databricks from Connect Cloud
CData Connect Cloud uses a straightforward, point-and-click interface to connect to data sources.
- Log into Connect Cloud, click Connections and click Add Connection
- Select "Databricks" from the Add Connection panel
-
Enter the necessary authentication properties to connect to Databricks.
To connect to a Databricks cluster, set the properties as described below.
Note: The needed values can be found in your Databricks instance by navigating to Clusters, and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.
- Server: Set to the Server Hostname of your Databricks cluster.
- HTTPPath: Set to the HTTP Path of your Databricks cluster.
- Token: Set to your personal access token (this value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).
- Click Create & Test
- Navigate to the Permissions tab in the Add Databricks Connection page and update the User-based permissions.
Configure Databricks Endpoints for Salesforce Connect
After connecting to Databricks, create a workspace and virtual dataset for your desired table(s).
- Navigate to the Virtual Datasets page and click Add to create a new Workspace (or select an existing workspace).
- Click Add to add new assets to the Workspace.
- Select the Databricks connection (e.g. Databricks1) and click Next.
- Select the table(s) you wish to work with and click Confirm.
- Make note of the OData Service URL for your workspace, e.g. https://cloud.cdata.com/api/odata/{workspace_name}
With the connection and Workspace configured, you are ready to connect to Databricks data from Salesforce Connect.
Connect to Databricks Data as an External Data Sources
Follow the steps below to connect to the feed produced by Connect Cloud.
- Log into Salesforce and click Setup -> Integrations -> External Data Sources.
- Click Now External Data Sources.
- Enter values for the following properties:
- External Data Sources: Enter a label to be used in list views and reports.
- Name: Enter a unique identifier.
- Type: Select the option "Salesforce Connect: OData 4.0".
- URL: Enter the URL to the OData endpoint of Connect Cloud: https://cloud.cdata.com/api/odata/{workspace_name}
- Select the Writable External Objects option.
- Select JSON in the Format menu.
- In the Authentication section, set the following properties:
- Identity Type: If all members of your organization will use the same credentials to access Connect Cloud, select "Named Principal". If the members of your organization will connect with their own credentials, select "Per User".
- Authentication Protocol: Select Password Authentication to use basic authentication.
- Certificate: Enter or browse to the certificate to be used to encrypt and authenticate communications from Salesforce to your server.
- Username: Enter a CData Connect Cloud username (e.g. [email protected].
- Password: Enter the user's PAT.
Synchronize Databricks Objects
After you have created the external data source, follow the steps below to create Databricks external objects that reflect any changes in the data source. You will synchronize the definitions for the Databricks external objects with the definitions for Databricks tables.
- Click the link for the external data source you created.
- Click Validate and Scan.
- Select the Databricks tables you want to work with as external objects.
Access Databricks Data as Salesforce Objects
After adding Databricks data as an external data source and syncing Databricks tables as external objects, you can use the external Databricks objects just as you would standard Salesforce objects.
-
Create a new tab with a filter list view:
-
Create reports of external objects:
-
Create, update, and delete Databricks objects from the Salesforce dashboard:
Simplified Access to Databricks Data from Cloud Applications
At this point, you have a direct, cloud-to-cloud connection to live Databricks data from Salesforce. For more information on gaining simplified access to data from more than 100 SaaS, Big Data, and NoSQL sources in cloud applications like Salesforce, refer to our Connect Cloud page.