Query Spark Data in DataGrip



Create a Data Source for Spark in DataGrip and use SQL to query live Spark data.

DataGrip is a database IDE that allows SQL developers to query, create, and manage databases. When paired with the CData JDBC Driver for Apache Spark, DataGrip can work with live Spark data. This article shows how to establish a connection to Spark data in DataGrip and use the table editor to load Spark data.

Create a New Driver Definition for Spark

The steps below describe how to create a new Data Source in DataGrip for Spark.

  1. In DataGrip, click File -> New > Project and name the project
  2. In the Database Explorer, click the plus icon () and select Driver.
  3. In the Driver tab:
    • Set Name to a user-friendly name (e.g. "CData Spark Driver")
    • Set Driver Files to the appropriate JAR file. To add the file, click the plus (), select "Add Files," navigate to the "lib" folder in the driver's installation directory and select the JAR file (e.g. cdata.jdbc.sparksql.jar).
    • Set Class to cdata.jdbc.sparksql.SparkSQL.jar
    Additionally, in the advanced tab you can change driver properties and some other settings like VM Options, VM environment, VM home path, DBMS, etc
    • For most cases, change the DBMS type to "Unknown" in Expert options to avoid native SQL Server queries (Transact-SQL), which might result in an invalid function error
  4. Click "Apply" then "OK" to save the Connection

Configure a Connection to Spark

  1. Once the connection is saved, click the plus (), then "Data Source" then "CData Spark Driver" to create a new Spark Data Source.
  2. In the new window, configure the connection to Spark with a JDBC URL.

    Built-in Connection String Designer

    For assistance in constructing the JDBC URL, use the connection string designer built into the Spark JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

    java -jar cdata.jdbc.sparksql.jar

    Fill in the connection properties and copy the connection string to the clipboard.

    Set the Server, Database, User, and Password connection properties to connect to SparkSQL.

  3. Set URL to the connection string, e.g., jdbc:sparksql:Server=127.0.0.1;
  4. Click "Apply" and "OK" to save the connection string

At this point, you will see the data source in the Data Explorer.

Execute SQL Queries Against Spark

To browse through the Spark entities (available as tables) accessible through the JDBC Driver, expand the Data Source.

To execute queries, right click on any table and select "New" -> "Query Console."

In the Console, write the SQL query you wish to execute. For example: SELECT City, Balance FROM Customers

Download a free, 30-day trial of the CData JDBC Driver for Apache Spark and start working with your live Spark data in DataGrip. Reach out to our Support Team if you have any questions.

Ready to get started?

Download a free trial of the Apache Spark Driver to get started:

 Download Now

Learn more:

Apache Spark Icon Apache Spark JDBC Driver

Rapidly create and deploy powerful Java applications that integrate with Apache Spark.