Load Phoenix Data to a Database Using Embulk

Ready to get started?

Download for a free trial:

Download Now

Learn more:

Phoenix JDBC Driver

Rapidly create and deploy powerful Java applications that integrate with HBase through Apache Phoenix.



Use CData JDBC Driver for Phoenix with the open source ETL/ELT tool Embulk to load Phoenix data to a database.

Embulk is an open source bulk data loader. When paired with the CData JDBC Driver for Phoenix, Embulk easily loads data from Phoenix to any supported destination. In this article, we explain how to use the CData JDBC Driver for Phoenix in Embulk to load Phoenix data to a MySQL dtabase.

With built-in optimized data processing, the CData JDBC Driver offers unmatched performance for interacting with live Phoenix data. When you issue complex SQL queries to Phoenix, the driver pushes supported SQL operations, like filters and aggregations, directly to Phoenix and utilizes the embedded SQL engine to process unsupported operations client-side (often SQL functions and JOIN operations).

Configure a JDBC Connection to Phoenix Data

Before creating a bulk load job in Embulk, note the installation location for the JAR file for the JDBC Driver (typically C:\Program Files\CData\CData JDBC Driver for Phoenix\lib).

Embulk supports JDBC connectivity, so you can easily connect to Phoenix and execute SQL queries. Before creating a bulk load job, create a JDBC URL for authenticating with Phoenix.

Connect to Apache Phoenix via the Phoenix Query Server. Set the Server and Port (if different from the default port) properties to connect to Apache Phoenix. The Server property will typically be the host name or IP address of the server hosting Apache Phoenix.

Authenticating to Apache Phoenix

By default, no authentication will be used (plain). If authentication is configured for your server, set AuthScheme to NEGOTIATE and set the User and Password properties (if necessary) to authenticate through Kerberos.

Built-in Connection String Designer

For assistance in constructing the JDBC URL, use the connection string designer built into the Phoenix JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

java -jar cdata.jdbc.apachephoenix.jar

Fill in the connection properties and copy the connection string to the clipboard.

Below is a typical JDBC connection string for Phoenix:

jdbc:apachephoenix:Server=localhost;Port=8765;

Load Phoenix Data in Embulk

After installing the CData JDBC Driver and creating a JDBC connection string, install the required Embulk plugins.

Install Embulk Input & Output Plugins

  1. Install the JDBC Input Plugin in Embulk.
    https://github.com/embulk/embulk-input-jdbc/tree/master/embulk-input-jdbc
  2. embulk gem install embulk-input-jdbc
  3. In this article, we use MySQL as the destination database. You can also choose SQL Server, PostgreSQL, or Google BigQuery as the destination using the output Plugins.
    https://github.com/embulk/embulk-output-jdbc/tree/master/embulk-output-mysql embulk gem install embulk-output-mysql

With the input and output plugins installed, we are ready to load Phoenix data into MySQL using Embulk.

Create a Job to Load Phoenix Data

Start by creating a config file in Embulk, using a name like apachephoenix-mysql.yml.

  1. For the input plugin options, use the CData JDBC Driver for Phoenix, including the path to the driver JAR file, the driver class (e.g. cdata.jdbc.apachephoenix.ApachePhoenixDriver), and the JDBC URL from above
  2. For the output plugin options, use the values and credentials for the MySQL database

Sample Config File (apachephoenix-mysql.yml)

in: type: jdbc driver_path: C:\Program Files\CData\CData JDBC Driver for Phoenix 20xx\lib\cdata.jdbc.apachephoenix.jar driver_class: cdata.jdbc.apachephoenix.ApachePhoenixDriver url: jdbc:apachephoenix:Server=localhost;Port=8765; table: "MyTable" out: type: mysql host: localhost database: DatabaseName user: UserId password: UserPassword table: "MyTable" mode: insert

After creating the file, run the Embulk job.

embulk run apachephoenix-mysql.yml

After running the the Embulk job, find the Salesforce data in the MySQL table.

Load Filtered Phoenix Data

In addition to loading data directly from a table, you can use a custom SQL query to have more granular control of the data loaded. You can also perform increment loads by setting a last updated column in a SQL WHERE clause in the query field.

in: type: jdbc driver_path: C:\Program Files\CData\CData JDBC Driver for Phoenix 20xx\lib\cdata.jdbc.apachephoenix.jar driver_class: cdata.jdbc.apachephoenix.ApachePhoenixDriver url: jdbc:apachephoenix:Server=localhost;Port=8765; query: "SELECT Id, Column1 FROM MyTable WHERE [RecordId] = 1" out: type: mysql host: localhost database: DatabaseName user: UserId password: UserPassword table: "MyTable" mode: insert

More Information & Free Trial

By using CData JDBC Driver for Phoenix as a connector, Embulk can integrate Phoenix data into your data load jobs. And with drivers for more than 200+ other enterprise sources, you can integrate any enterprise SaaS, big data, or NoSQL source as well. Download a 30-day free trial and get started today.