Ready to get started?

Learn more about the CData JDBC Driver for HDFS or download a free trial:

Download Now

ETL HDFS in Oracle Data Integrator

This article shows how to transfer HDFS data into a data warehouse using Oracle Data Integrator.

Leverage existing skills by using the JDBC standard to connect to HDFS: Through drop-in integration into ETL tools like Oracle Data Integrator (ODI), the CData JDBC Driver for HDFS connects real-time HDFS data to your data warehouse, business intelligence, and Big Data technologies.

JDBC connectivity enables you to work with HDFS just as you would any other database in ODI. As with an RDBMS, you can use the driver to connect directly to the HDFS APIs in real time instead of working with flat files.

This article walks through a JDBC-based ETL -- HDFS to Oracle. After reverse engineering a data model of HDFS entities, you will create a mapping and select a data loading strategy -- since the driver supports SQL-92, this last step can easily be accomplished by selecting the built-in SQL to SQL Loading Knowledge Module.

Install the Driver

To install the driver, copy the driver JAR and .lic file, located in the installation folder, into the ODI userlib directory:

On Unix: ~/.odi/oracledi/userlib On Windows: %APPDATA%\Roaming\odi\oracledi\userlib

Restart ODI to complete the installation.

Reverse Engineer a Model

Reverse engineering the model retrieves metadata about the driver's relational view of HDFS data. After reverse engineering, you can query real-time HDFS data and create mappings based on HDFS tables.

  1. In ODI, connect to your repository and click New -> Model and Topology Objects.
  2. On the Model screen of the resulting dialog, enter the following information:
    • Name: Enter HDFS.
    • Technology: Select Generic SQL (for ODI Version 12.2+, select Microsoft SQL Server).
    • Logical Schema: Enter HDFS.
    • Context: Select Global.
  3. On the Data Server screen of the resulting dialog, enter the following information:
    • Name: Enter HDFS.
    • Driver List: Select Oracle JDBC Driver.
    • Driver: Enter cdata.jdbc.hdfs.HDFSDriver
    • URL: Enter the JDBC URL containing the connection string.

      In order to authenticate, set the following connection properties:

      • Host: Set this value to the host of your HDFS installation.
      • Port: Set this value to the port of your HDFS installation. Default port: 50070

      Built-in Connection String Designer

      For assistance in constructing the JDBC URL, use the connection string designer built into the HDFS JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

      java -jar cdata.jdbc.hdfs.jar

      Fill in the connection properties and copy the connection string to the clipboard.

      Below is a typical connection string:

      jdbc:hdfs:Host=sandbox-hdp.hortonworks.com;Port=50070;Path=/user/root;User=root;
  4. On the Physical Schema screen, enter the following information:
    • Schema (Schema): Enter HDFS.
    • Schema (Work Schema): Enter HDFS.
  5. In the opened model click Reverse Engineer to retrieve the metadata for HDFS tables.

Edit and Save HDFS Data

After reverse engineering you can now work with HDFS data in ODI. To view HDFS data, expand the Models accordion in the Designer navigator, right-click a table, and click View data.

Create an ETL Project

Follow the steps below to create an ETL from HDFS. You will load Files entities into the sample data warehouse included in the ODI Getting Started VM.

  1. Open SQL Developer and connect to your Oracle database. Right-click the node for your database in the Connections pane and click new SQL Worksheet.

    Alternatively you can use SQLPlus. From a command prompt enter the following:

    sqlplus / as sysdba
  2. Enter the following query to create a new target table in the sample data warehouse, which is in the ODI_DEMO schema. The following query defines a few columns that match the Files table in HDFS: CREATE TABLE ODI_DEMO.TRG_FILES (CHILDRENNUM NUMBER(20,0),FileId VARCHAR2(255));
  3. In ODI expand the Models accordion in the Designer navigator and double-click the Sales Administration node in the ODI_DEMO folder. The model is opened in the Model Editor.
  4. Click Reverse Engineer. The TRG_FILES table is added to the model.
  5. Right-click the Mappings node in your project and click New Mapping. Enter a name for the mapping and clear the Create Empty Dataset option. The Mapping Editor is displayed.
  6. Drag the TRG_FILES table from the Sales Administration model onto the mapping.
  7. Drag the Files table from the HDFS model onto the mapping.
  8. Click the source connector point and drag to the target connector point. The Attribute Matching dialog is displayed. For this example, use the default options. The target expressions are then displayed in the properties for the target columns.
  9. Open the Physical tab of the Mapping Editor and click FILES_AP in TARGET_GROUP.
  10. In the FILES_AP properties, select LKM SQL to SQL (Built-In) on the Loading Knowledge Module tab.

You can then run the mapping to load HDFS data into Oracle.