Ready to get started?

Learn more about the CData JDBC Driver for DataRobot or download a free trial:

Download Now

Connect to DataRobot Data in CloverETL

Transfer DataRobot data using the visual workflow in the CloverETL data integration tool.

The CData JDBC Driver for DataRobot enables you to use the data transformation components in CloverETL to execute queries to DataRobot data. In this article, you will use the JDBC Driver for DataRobot to set up a simple transfer into a flat file.

Connect to DataRobot as a JDBC Data Source

  1. Create the connection to DataRobot data. In a new CloverETL graph, right-click the Connections node in the Outline pane and click Connections -> Create Connection. The Database Connection wizard is displayed.
  2. Click the plus icon to load a driver from a JAR. Browse to the lib subfolder of the installation directory and select the cdata.jdbc.datarobot.jar file.
  3. Enter the JDBC URL. A typical JDBC URL is below: jdbc:datarobot:PredictionInstance=myinstance.orm.datarobot.com;DataFile=PATH\TO\input_file.csv;DataRobotKey=123-abc-456-def;User=username;Password=password;

    To connect to DataRobot, the following connection properties are required: User, Password, and PredictionInstance. DataRobotKey may also be required depending on your type of DataRobot predictions instance. If using the Predictions API, DataFile is required. The CSV DataFile should include a header row as the first row of the datafile. APIKey is not required, but can be supplied. If not supplied, the driver will handle obtaining an APIKey.

    User, DataRobotKey, and APIKey are the credentials for the DataRobot account.

    ProjectID, DataFile, and ModelId are the parameters for the project, dataset, and model type.

Query DataRobot Data with the DBInputTable Component

  1. Drag a DBInputTable from the Readers selection of the Palette onto the job flow and double-click it to open the configuration editor.
  2. In the DB connection property, select the DataRobot JDBC data source from the drop-down menu.
  3. Enter the SQL query. For example: SELECT * FROM Predictions

Write the Output of the Query to a UniversalDataWriter

  1. Drag a UniversalDataWriter from the Writers selection onto the jobflow.
  2. Double-click the UniversalDataWriter to open the configuration editor and add a file URL.
  3. Right-click the DBInputTable and then click Extract Metadata.
  4. Connect the output port of the DBInputTable to the UniversalDataWriter.
  5. In the resulting Select Metadata menu for the UniversalDataWriter, choose the Predictions table. (You can also open this menu by right-clicking the input port for the UniversalDataWriter.)
  6. Click Run to write to the file.
 
 
ダウンロード