Ready to get started?

Learn more about the CData JDBC Driver for Apache Hive or download a free trial:

Download Now

A PostgreSQL Interface for Hive Data

Use the Remoting features of the Hive JDBC Driver to create a PostgreSQL entry-point for data access.

There are a vast number of PostgreSQL clients available on the Internet. From standard Drivers to BI and Analytics tools, PostgreSQL is a popular interface for data access. Using our JDBC Drivers, you can now create PostgreSQL entry-points that you can connect to from any standard client.

To access Hive data as a PostgreSQL database, use the CData JDBC Driver for Hive and a JDBC foreign data wrapper (FDW). In this article, we compile the FDW, install it, and query Hive data from PostgreSQL Server.

Connect to Hive Data as a JDBC Data Source

To connect to Hive as a JDBC data source, you will need the following:

  • Driver JAR path: The JAR is located in the lib subfolder of the installation directory.
  • Driver class: cdata.jdbc.apachehive.ApacheHiveDriver

  • JDBC URL: The URL must start with "jdbc:apachehive:" and can include any of the connection properties in name-value pairs separated with semicolons. Set the Server, Port, TransportMode, and AuthScheme connection properties to connect to Hive.

    Built-in Connection String Designer

    For assistance in constructing the JDBC URL, use the connection string designer built into the Hive JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

    java -jar cdata.jdbc.apachehive.jar

    Fill in the connection properties and copy the connection string to the clipboard.

    A typical JDBC URL is below:

    jdbc:apachehive:Server=127.0.0.1;Port=10000;TransportMode=BINARY;

Build the JDBC Foreign Data Wrapper

The Foreign Data Wrapper can be installed as an extension to PostgreSQL, without recompiling PostgreSQL. The jdbc2_fdw extension is used as an example.

  1. Add a symlink from the shared object for your version of the JRE to /usr/lib/libjvm.so. For example: ln -s /usr/lib/jvm/java-6-openjdk/jre/lib/amd64/server/libjvm.so /usr/lib/libjvm.so
  2. Start the build: make install USE_PGXS=1

Query Hive Data as a PostgreSQL Database

After you have installed the extension, follow the steps below to start executing queries to Hive data:

  1. Log into your database.
  2. Load the extension for the database: CREATE EXTENSION jdbc2_fdw;
  3. Create a server object for Hive: CREATE SERVER ApacheHive FOREIGN DATA WRAPPER jdbc2_fdw OPTIONS ( drivername 'cdata.jdbc.apachehive.ApacheHiveDriver', url 'jdbc:apachehive:Server=127.0.0.1;Port=10000;TransportMode=BINARY;', querytimeout '15', jarfile '/home/MyUser/CData/CData\ JDBC\ Driver\ for\ Salesforce MyDriverEdition/lib/cdata.jdbc.apachehive.jar');
  4. Create a user mapping for the username and password of a user known to the MySQL daemon. CREATE USER MAPPING for postgres SERVER ApacheHive OPTIONS ( username 'admin', password 'test');
  5. Create a foreign table in your local database: postgres=# CREATE FOREIGN TABLE customers ( customers_id text, customers_City text, customers_CompanyName numeric) SERVER ApacheHive OPTIONS ( table_name 'customers');
You can now execute read/write commands to Hive: postgres=# SELECT * FROM customers;