Create a Kafka Data Source for Denodo Platform

詳細情報をご希望ですか?

無償トライアル:

ダウンロードへ

製品の詳細情報へ:

Apache Kafka JDBC Driver

Apache Kafka データに連携するJava アプリケーションを素早く、簡単に開発できる便利なドライバー。



Use the CData JDBC Driver for Kafka to create a virtual data source for Kafka data in the Denodo Virtual DataPort Administrator.

Denodo Platform is a data virtualization product providing a single point of contact for enterprise database data. When paired with the CData JDBC Driver for Kafka, Denodo users can work with live Kafka data alongside other enterprise data sources. This article walks through creating a virtual data source for Kafka in the Denodo Virtual DataPort Administrator.

With built-in optimized data processing, the CData JDBC Driver offers unmatched performance for interacting with live Kafka data. When you issue complex SQL queries to Kafka, the driver pushes supported SQL operations, like filters and aggregations, directly to Kafka and utilizes the embedded SQL engine to process unsupported operations client-side (often SQL functions and JOIN operations). Its built-in dynamic metadata querying allows you to work with and analyze Kafka data using native data types.

Create a Kafka Virtual Port

To connect to live Kafka data from Denodo, you need to copy the JDBC Driver JAR file to the external library directory for Denodo and create a new JDBC Data Source from the Virtual DataPort Administrator tool.

  1. Download the CData JDBC Driver for Kafka installer, unzip the package, and run the JAR file to install the driver.
  2. Copy the JAR File (and license file if it exists) from the installation location (typically C:\Program Files\CData\CData JDBC Driver for Kafka\lib\) to the Denodo external library directory (C:\Denodo\Denodo Platform\lib-external\jdbc-drivers\cdata-apachekafka-19).
  3. Open the Denodo Virtual DataPort Administrator tool and navigate to the Server Explorer tab.
  4. Right-click "admin" and select New -> Data source -> JDBC.
  5. Configure the JDBC Connection:
    • Name: your choice, e.g.: apachekafka
    • Database adapter: Generic
    • Driver class path: C:\Denodo\Denodo Platform\lib-external\jdbc-drivers\cdata-apachekafka-19
    • Driver class: cdata.jdbc.apachekafka.ApacheKafkaDriver
    • Database URI: Set this to a JDBC URL using the necessary connection properties. For example,

      jdbc:apachekafka:User=admin;Password=pass;BootStrapServers=https://localhost:9091;Topic=MyTopic;

      Information on creating the Database URI follows:

      Built-In Connection String Designer

      For assistance in constructing the JDBC URL, use the connection string designer built into the Kafka JDBC Driver. Either double-click the JAR file or execute the jar file from the command-line.

      java -jar cdata.jdbc.apachekafka.jar

      Fill in the connection properties and copy the connection string to the clipboard.

      BootstrapServers およびTopic プロパティを設定して、Apache Kafka サーバーのアドレスと、対話するトピックを指定します。

      認可メカニズム

      • SASL PlainUser およびPassword プロパティを指定する必要があります。AuthScheme は'Plain' に設定します。
      • SASL SSLUser およびPassword プロパティを指定する必要があります。AuthScheme は'Scram' に、UseSSL はtrue に設定します。
      • SSLSSLCert およびSSLCertPassword プロパティを指定する必要があります。UseSSL はtrue に設定します。
      • KerberosUser およびPassword プロパティを指定する必要があります。AuthScheme は'Kerberos' に設定します。

      サーバー証明書を信頼する必要がある場合があります。そのような場合は、必要に応じてTrustStorePath およびTrustStorePassword を指定してください。

  6. Click the "Test connection" button to confirm the configuration and click Save.

View Kafka Data in the VirtualPort Administrator Tool

After creating the data source, you can create a base view of Kafka data for use in the Denodo Platform.

  1. Click the "Create base view" button in the newly created VirtualPort (admin.ApacheKafka).
  2. Expand the object tree and select the objects (tables) you wish to import.
  3. Click the "Create selected" button to create views of the Kafka data.
    Optional: Click "Create associations from foreign keys" to define relationships between the objects.
  4. With the view(s) created, navigate to a table (cdata_apachekafka_sampletable_1) in the Server Explorer and double-click the selected table.
  5. In the new tab, click "Execution panel" to open a query panel.
  6. Customize the query in the "Execute" tab or use the default:
    SELECT * FROM cdata_apachekafka_sampletable_1 CONTEXT ('i18n'='us_est', 'cache_wait_for_load'='true')
    
  7. Click Execute to view the data.

With the base view created, you can now work with live Kafka data like you would any other data source in Denodo Platform, for example, querying Kafka in the Denodo Data Catalog.

Download a free, 30-day trial of the CData JDBC Driver for Kafka and start working with your live Kafka data in Denodo Platform. Reach out to our Support Team if you have any questions.