Integrate Elasticsearch Data in Pentaho Data Integration



Build ETL pipelines based on Elasticsearch data in the Pentaho Data Integration tool.

The CData JDBC Driver for Elasticsearch enables access to live data from data pipelines. Pentaho Data Integration is an Extraction, Transformation, and Loading (ETL) engine that data, cleanses the data, and stores data using a uniform format that is accessible.This article shows how to connect to Elasticsearch data as a JDBC data source and build jobs and transformations based on Elasticsearch data in Pentaho Data Integration.

About Elasticsearch Data Integration

Accessing and integrating live data from Elasticsearch has never been easier with CData. Customers rely on CData connectivity to:

  • Access both the SQL endpoints and REST endpoints, optimizing connectivity and offering more options when it comes to reading and writing Elasticsearch data.
  • Connect to virtually every Elasticsearch instance starting with v2.2 and Open Source Elasticsearch subscriptions.
  • Always receive a relevance score for the query results without explicitly requiring the SCORE() function, simplifying access from 3rd party tools and easily seeing how the query results rank in text relevance.
  • Search through multiple indices, relying on Elasticsearch to manage and process the query and results instead of the client machine.

Users frequently integrate Elasticsearch data with analytics tools such as Crystal Reports, Power BI, and Excel, and leverage our tools to enable a single, federated access layer to all of their data sources, including Elasticsearch.

For more information on CData's Elasticsearch solutions, check out our Knowledge Base article: CData Elasticsearch Driver Features & Differentiators.


Getting Started


Configure to Elasticsearch Connectivity

Set the Server and Port connection properties to connect. To authenticate, set the User and Password properties, PKI (public key infrastructure) properties, or both. To use PKI, set the SSLClientCert, SSLClientCertType, SSLClientCertSubject, and SSLClientCertPassword properties.

The data provider uses X-Pack Security for TLS/SSL and authentication. To connect over TLS/SSL, prefix the Server value with 'https://'. Note: TLS/SSL and client authentication must be enabled on X-Pack to use PKI.

Once the data provider is connected, X-Pack will then perform user authentication and grant role permissions based on the realms you have configured.

Built-in Connection String Designer

For assistance in constructing the JDBC URL, use the connection string designer built into the Elasticsearch JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

java -jar cdata.jdbc.elasticsearch.jar

Fill in the connection properties and copy the connection string to the clipboard.

When you configure the JDBC URL, you may also want to set the Max Rows connection property. This will limit the number of rows returned, which is especially helpful for improving performance when designing reports and visualizations.

Below is a typical JDBC URL:

jdbc:elasticsearch:Server=127.0.0.1;Port=9200;User=admin;Password=123456;

Save your connection string for use in Pentaho Data Integration.

Connect to Elasticsearch from Pentaho DI

Open Pentaho Data Integration and select "Database Connection" to configure a connection to the CData JDBC Driver for Elasticsearch

  1. Click "General"
  2. Set Connection name (e.g. Elasticsearch Connection)
  3. Set Connection type to "Generic database"
  4. Set Access to "Native (JDBC)"
  5. Set Custom connection URL to your Elasticsearch connection string (e.g.
    jdbc:elasticsearch:Server=127.0.0.1;Port=9200;User=admin;Password=123456;
  6. Set Custom driver class name to "cdata.jdbc.elasticsearch.ElasticsearchDriver"
  7. Test the connection and click "OK" to save.

Create a Data Pipeline for Elasticsearch

Once the connection to Elasticsearch is configured using the CData JDBC Driver, you are ready to create a new transformation or job.

  1. Click "File" >> "New" >> "Transformation/job"
  2. Drag a "Table input" object into the workflow panel and select your Elasticsearch connection.
  3. Click "Get SQL select statement" and use the Database Explorer to view the available tables and views.
  4. Select a table and optionally preview the data for verification.

At this point, you can continue your transformation or jb by selecting a suitable destination and adding any transformations to modify, filter, or otherwise alter the data during replication.

Free Trial & More Information

Download a free, 30-day trial of the CData JDBC Driver for Elasticsearch and start working with your live Elasticsearch data in Pentaho Data Integration today.

Ready to get started?

Download a free trial of the Elasticsearch Driver to get started:

 Download Now

Learn more:

Elasticsearch Icon Elasticsearch JDBC Driver

Rapidly create and deploy powerful Java applications that integrate with Elasticsearch.