Consume Azure Data Lake Storage OData Feeds in SAP Lumira

Ready to get started?

Learn more or sign up for a free trial:

CData API Server



Use the API Server to create data visualizations on Azure Data Lake Storage feeds that reflect any changes in SAP Lumira.

You can use the CData API Server and the ADO.NET Provider for Azure Data Lake Storage (or any of 200+ other ADO.NET Providers) to create data visualizations based on Azure Data Lake Storage data in SAP Lumira. The API Server enables connectivity to live data: dashboards and reports can be refreshed on demand. This article shows how to create a chart that is always up to date.

Set Up the API Server

Follow the steps below to begin producing secure Azure Data Lake Storage OData services:

Deploy

The API Server runs on your own server. On Windows, you can deploy using the stand-alone server or IIS. On a Java servlet container, drop in the API Server WAR file. See the help documentation for more information and how-tos.

The API Server is also easy to deploy on Microsoft Azure, Amazon EC2, and Heroku.

Connect to Azure Data Lake Storage

After you deploy the API Server and the ADO.NET Provider for Azure Data Lake Storage, provide authentication values and other connection properties needed to connect to Azure Data Lake Storage by clicking Settings -> Connection and adding a new connection in the API Server administration console.

Authenticating to a Gen 1 DataLakeStore Account

Gen 1 uses OAuth 2.0 in Azure AD for authentication.

For this, an Active Directory web application is required. You can create one as follows:

  1. Sign in to your Azure Account through the .
  2. Select "Azure Active Directory".
  3. Select "App registrations".
  4. Select "New application registration".
  5. Provide a name and URL for the application. Select Web app for the type of application you want to create.
  6. Select "Required permissions" and change the required permissions for this app. At a minimum, "Azure Data Lake" and "Windows Azure Service Management API" are required.
  7. Select "Key" and generate a new key. Add a description, a duration, and take note of the generated key. You won't be able to see it again.

To authenticate against a Gen 1 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen1.
  • Account: Set this to the name of the account.
  • OAuthClientId: Set this to the application Id of the app you created.
  • OAuthClientSecret: Set this to the key generated for the app you created.
  • TenantId: Set this to the tenant Id. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

Authenticating to a Gen 2 DataLakeStore Account

To authenticate against a Gen 2 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen2.
  • Account: Set this to the name of the account.
  • FileSystem: Set this to the file system which will be used for this account.
  • AccessKey: Set this to the access key which will be used to authenticate the calls to the API. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

When you configure the connection, you may also want to set the Max Rows connection property. This will limit the number of rows returned, which is especially helpful for improving performance when designing reports and visualizations.

You can then choose the Azure Data Lake Storage entities you want to allow the API Server access to by clicking Settings -> Resources.

Authorize API Server Users

After determining the OData services you want to produce, authorize users by clicking Settings -> Users. The API Server uses authtoken-based authentication and supports the major authentication schemes. Access can also be restricted based on IP address; by default, only connections to the local machine are allowed. You can authenticate as well as encrypt connections with SSL.

Connect to Azure Data Lake Storage from SAP Lumira

Follow the steps below to retrieve Azure Data Lake Storage data into SAP Lumira. You can execute an SQL query or use the UI.

  1. In SAP Lumira, click File -> New -> Query with SQL. The Add New Dataset dialog is displayed.
  2. Expand the Generic section and click the Generic OData 2.0 Connector option.
  3. In the Service Root URI box, enter the OData endpoint of the API Server. This URL will resemble the following: https://your-server:8032/api.rsc
  4. In the User Name and Password boxes, enter the username and authtoken of an API user. These credentials will be used in HTTP Basic authentication.
  5. Select entities in the tree or enter an SQL query. This article imports Azure Data Lake Storage Resources entities.

  6. When you click Connect, SAP Lumira will generate the corresponding OData request and load the results into memory. You can then use any of the data processing tools available in SAP Lumira, such as filters, aggregates, and summary functions.

Create Data Visualizations

After you have imported the data, you can create data visualizations in the Visualize room. Follow the steps below to create a basic chart.

  1. In the Measures and Dimensions pane, drag measures and dimensions onto the x-axis and y-axis fields in the Visualization Tools pane. SAP Lumira automatically detects dimensions and measures from the metadata service of the API Server.

    By default, the SUM function is applied to all measures. Click the gear icon next to a measure to change the default summary.

  2. In the Visualization Tools pane, select the chart type.
  3. In the Chart Canvas pane, apply filters, sort by measures, add rankings, and update the chart with the current Azure Data Lake Storage data.