We are proud to share our inclusion in the 2024 Gartner Magic Quadrant for Data Integration Tools. We believe this recognition reflects the differentiated business outcomes CData delivers to our customers.
Get the Report →Build Semantic Layer Views for Spark Data in APOS Live Data Gateway
Use the CData Connector for Spark in APOS Live Data Gateway to build Semantic Layer Views for Spark data.
APOS Live Data Gateway (LDG) is a data connection and data transformation solution that enables live data connectivity and expanded data source options for SAP Analytics Cloud and other SAP solutions. When paired with CData Connectors, users can build semantic layer views for live Spark data, enabling real-time analytics on Spark just like working with a relational database.
With built-in optimized data processing, the CData Connector offers unmatched performance for interacting with live Spark data. When you issue complex SQL queries to Spark, the driver pushes supported SQL operations, like filters and aggregations, directly to Spark and utilizes the embedded SQL engine to process unsupported operations client-side (often SQL functions and JOIN operations). Its built-in dynamic metadata querying allows you to work with and analyze Spark data using native data types.
Download and Install the Connector Files
In order to access Spark data through the APOS Live Data Gateway, you will need to download the connector files from APOS and install them on the machine hosting the Live Data Gateway. An APOS representative can deliver the necessary files.
- Install the CData DLL file (System.Data.CData.SparkSQL.dll) to the APOS Live Data Gateway installation directory (C:\Program Files\Live Data Gateway\Admin\ by default).
- Install the CData JAR file (cdata.jdbc.sparksql.jar) to the ConnectionTest_lib folder in the installation directory (C:\Program Files\Live Data Gateway\Admin\ConnectionTest_lib\ by default).
- Install the CData JAR file (cdata.jdbc.sparksql.jar) to the lib folder in the Web UI installation directory (e.g.: C:\LDG_WebUI\lib\)
Configuring the Spark Connection String
Before establishing the connection to Spark from the APOS Live Data Gateway, you need to configure the Spark JDBC Connection String.
Built-in Connection String Designer
For assistance in constructing the JDBC Connection String, use the connection string designer built into the Spark JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.
java -jar cdata.jdbc.sparksql.jar
Fill in the connection properties and copy the connection string to the clipboard.
Set the Server, Database, User, and Password connection properties to connect to SparkSQL.
Your connection string will look similar to the following:
jdbc:SparkSQL:Server=127.0.0.1;
Connecting to Spark & Creating a Semantic Layer View
After installing the connector files and configuring the connection string, you are ready to connect to Spark in the Live Data Gateway Admin tool and build a semantic layer view in the Live Data Gateway Web UI.
Configuring the Connection to Spark
- Log into your APOS Live Data Gateway Manager
- If you haven't already, update your APOS LDG license file
- Click File -> Configurations
- Click on the "..." Menu for the License
- Select the license file from the APOS team that includes your CData Connector license
- In the APOS Live Data Gateway Manager, click "Add"
- In the APOS Live Data Gateway On the Connection tab, configure the connection:
- Set Data Source to "Database"
- Set Database to "CData Drivers"
- Set Database CData Drivers to "Spark"
- Set Connection String to the connection string configured earlier (e.g.:
jdbc:SparkSQL:Server=127.0.0.1;
- Set Driver Class to "cdata.jdbc.sparksql.SparkSQLDriver" (this should be set by default)
- Click Test Connection
- Click Save
- Give your connection a unique prefix (e.g. "sparksql")
- Highlight the newly created connection and click File -> "Approve Users For Web UI"
- Approve the appropriate DB users to create views and click "Save"
At this point, we are ready to build our semantic layer view in the Live Data Gateway Web UI.
Creating a Semantic Layer View
- In your browser, navigate to the APOS Live Data Gateway Portal
- Select a Connection (e.g. "sparksql")
- Since Spark does not require a User or Password to authenticate, you may use whatever values you wish for User Name and Password
- Click "Login"
- Once connected, click "Semantic Layer" to create a new semantic layer view
- Click "New Semantic Layer View"
- Set the Semantic Layer View Prefix and Semantic Layer View Name
- Click "Step 2"
- Select the table(s) and column(s) you wish to include in your view
- Click "Step 3"
- Select the Measures from the available table columns
- Click "Step 5" (we skipped the "Extra Dimensions" step)
- Add any Variable Prompts
- Click "Step 6"
- Define any Table Joins
- Click "Review"
- Review you semantic layer view and click "Save"
With the Semantic Layer View created, you are ready to access your Spark data through the APOS Live Data Gateway, enabling real-time data connectivity to Spark data from SAP Analytics Cloud and other SAP solutions.
More Information & Free Evaluation
Please visit APOS Systems - APOS Solutions - Request Evaluation Software to request evaluation software or email [email protected] for more information on working with your live Spark data in APOS Live Data Gateway.