Integrate SAS Data Sets Data in Pentaho Data Integration



Build ETL pipelines based on SAS Data Sets data in the Pentaho Data Integration tool.

The CData JDBC Driver for SAS Data Sets enables access to live data from data pipelines. Pentaho Data Integration is an Extraction, Transformation, and Loading (ETL) engine that data, cleanses the data, and stores data using a uniform format that is accessible.This article shows how to connect to SAS Data Sets data as a JDBC data source and build jobs and transformations based on SAS Data Sets data in Pentaho Data Integration.

Configure to SAS Data Sets Connectivity

Set the following connection properties to connect to your SAS DataSet files:

Connecting to Local Files

  • Set the Connection Type to "Local." Local files support SELECT, INSERT, and DELETE commands.
  • Set the URI to a folder containing SAS files, e.g. C:\PATH\TO\FOLDER\.

Connecting to Cloud-Hosted SAS DataSet Files

While the driver is capable of pulling data from SAS DataSet files hosted on a variety of cloud data stores, INSERT, UPDATE, and DELETE are not supported outside of local files in this driver.

Set the Connection Type to the service hosting your SAS DataSet files. A unique prefix at the beginning of the URI connection property is used to identify the cloud data store and the remainder of the path is a relative path to the desired folder (one table per file) or single file (a single table). For more information, refer to the Getting Started section of the Help documentation.

Built-in Connection String Designer

For assistance in constructing the JDBC URL, use the connection string designer built into the SAS Data Sets JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

java -jar cdata.jdbc.sasdatasets.jar

Fill in the connection properties and copy the connection string to the clipboard.

When you configure the JDBC URL, you may also want to set the Max Rows connection property. This will limit the number of rows returned, which is especially helpful for improving performance when designing reports and visualizations.

Below is a typical JDBC URL:

jdbc:sasdatasets:URI=C:/myfolder;

Save your connection string for use in Pentaho Data Integration.

Connect to SAS Data Sets from Pentaho DI

Open Pentaho Data Integration and select "Database Connection" to configure a connection to the CData JDBC Driver for SAS Data Sets

  1. Click "General"
  2. Set Connection name (e.g. SAS Data Sets Connection)
  3. Set Connection type to "Generic database"
  4. Set Access to "Native (JDBC)"
  5. Set Custom connection URL to your SAS Data Sets connection string (e.g.
    jdbc:sasdatasets:URI=C:/myfolder;
  6. Set Custom driver class name to "cdata.jdbc.sasdatasets.SASDataSetsDriver"
  7. Test the connection and click "OK" to save.

Create a Data Pipeline for SAS Data Sets

Once the connection to SAS Data Sets is configured using the CData JDBC Driver, you are ready to create a new transformation or job.

  1. Click "File" >> "New" >> "Transformation/job"
  2. Drag a "Table input" object into the workflow panel and select your SAS Data Sets connection.
  3. Click "Get SQL select statement" and use the Database Explorer to view the available tables and views.
  4. Select a table and optionally preview the data for verification.

At this point, you can continue your transformation or jb by selecting a suitable destination and adding any transformations to modify, filter, or otherwise alter the data during replication.

Free Trial & More Information

Download a free, 30-day trial of the CData JDBC Driver for SAS Data Sets and start working with your live SAS Data Sets data in Pentaho Data Integration today.

Ready to get started?

Download a free trial of the SAS Data Sets Driver to get started:

 Download Now

Learn more:

SAS Data Sets Icon SAS Data Sets JDBC Driver

Rapidly create and deploy powerful Java applications that integrate with SAS Data Sets.