Build Databricks-Powered Applications in Cursor with CData Code Assist MCP
Cursor is an AI-powered code editor that integrates agentic AI into everyday development workflows. With support for MCP, Cursor can connect to local tools and enterprise data sources directly from the editor, enabling natural language interaction with live systems without switching context.
Model Context Protocol (MCP) is an open standard for connecting LLM clients to external services through structured tool interfaces. MCP servers expose capabilities such as schema discovery and live querying, allowing AI agents to retrieve and reason over real-time data safely and consistently.
In this article, we guide you through installing the CData Code Assist MCP for Databricks, configuring the connection to Databricks, connecting the Code Assist MCP add-on to Cursor, and querying live Databricks data from within the editor.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Step 1: Download and install the CData Code Assist MCP for Databricks
- To begin, download the CData Code Assist MCP for Databricks
- Find and double-click the installer to begin the installation
- Run the installer and follow the prompts to complete the installation
When the installation is complete, you are ready to configure your Code Assist MCP add-on by connecting to Databricks.
Step 2: Configure the connection to Databricks
- After installation, open the CData Code Assist MCP for Databricks configuration wizard
NOTE: If the wizard does not open automatically, search for "CData Code Assist MCP for Databricks" in the Windows search bar and open the application.
- In MCP Configuration > Configuration Name, either select an existing configuration or choose
to create a new one
- Name the configuration (e.g. "cdata_databricks") and click OK
-
Enter the appropriate connection properties in the configuration wizard
To connect to a Databricks cluster, set the properties as described below.
Note: The needed values can be found in your Databricks instance by navigating to Clusters, and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.
- Server: Set to the Server Hostname of your Databricks cluster.
- HTTPPath: Set to the HTTP Path of your Databricks cluster.
- Token: Set to your personal access token (this value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).
- Click Connect to authenticate with Databricks
- Then, click Save Configuration to save the Code Assist MCP add-on
This process creates a .mcp configuration file that Cursor will reference when launching the Code Assist MCP add-on. Now with your Code Assist MCP add-on configured, you are ready to connect it to Cursor.
Step 3: Connect the Code Assist MCP add-on to Cursor
- Download the Cursor desktop application and complete the sign-up flow for your account
-
From the top menu, click Settings to open the settings panel
-
In the left navigation, open the Tools & MCP tab and click Add Custom MCP
- Cursor opens an mcp.json file in the editor
- Add the code shown below and save the file
- After saving and testing your connection in the configuration wizard, click Next
- Select Cursor from the AI MCP Tool dropdown
- Follow the MCP Client Instructions to create the required folders for the MCP config
- Copy the displayed JSON code and paste it into your configuration file
- In Cursor, open the project folder you created with the mcp.json config
- The Code Assist MCP add-on should appear as Running under Installed MCP Servers
Option 1: Manually add the MCP configuration
{
"mcpServers": {
"cdata-local": {
"command": "C:/Program Files/Java/jdk-17/bin/java.exe",
"args": [
"-jar",
"C:/Program Files/CData/CData Code Assist MCP for Databricks/lib/cdata.mcp.databricks.jar",
"cdata_databricks"
]
}
}
}
NOTE: The command value should point to your Java 17+ java.exe executable, and the JAR path should point to the installed CData Code Assist MCP add-on .jar file. The final argument must match the MCP configuration name you saved in the CData configuration wizard (e.g. "cdata_databricks").
Option 2: Copy the MCP configuration from the CData Code Assist MCP for Databricks UI
Step 4: Query live Databricks data in Cursor
- From the top bar, click Toggle AI Pane to open the chat window
- Ask questions about your Databricks data using natural language. For example:
"List all tables available in my Databricks data connection."
Cursor is now fully integrated with CData Code Assist MCP for Databricks and can use the MCP tools exposed to explore schemas and execute live queries against Databricks.
Build with Code Assist MCP. Deploy with CData Drivers.
Download Code Assist MCP for free and give your AI tools schema-aware access to live Databricks data during development. When you're ready to move to production, CData Databricks Drivers deliver the same SQL-based access with enterprise-grade performance, security, and reliability.
Visit the CData Community to share insights, ask questions, and explore what's possible with MCP-powered AI workflows.