今すぐお試しください!

製品の詳細CData SSIS Components for Apache Spark を確認して、無償評価版をダウンロード:

今すぐダウンロード

Export Data from SQL Server to Spark through SSIS

Easily push SQL Server data to Spark using the CData SSIS Tasks for Spark.

SQL Server databases are commonly used to store enterprise records. It is often necessary to move this data to other locations. The CData SSIS Task for Spark allows you to easily transfer Spark data. In this article you will export data from SQL Server to Spark.

Add Source and Destination Components

To get started, add a new ADO.NET Source control and a new Spark Destination control to the data flow task.

Configure the ADO.NET Source

Follow the steps below to specify properties required to connect to the SQL Server instance.

  1. Open the ADO.NET Source and add a new connection. Enter your server and database information here.
  2. In the Data access mode menu, select "Table or view" and select the table or view to export into Spark.
  3. Close the ADO NET Source wizard and connect it to the destination component.

Create a New Connection Manager for Spark

Follow the steps below to set required connection properties in the Connection Manager.

  1. Create a new connection manager: In the Connection Manager window, right-click and then click New Connection. The Add SSIS Connection Manager dialog is displayed.
  2. Select CData SparkSQL Connection Manager in the menu.
  3. Configure the connection properties. A typical connection requires:
    • Server

    SparkSQL への接続

    SparkSQL への接続を確立するには以下を指定します。

    • Server:SparkSQL をホストするサーバーのホスト名またはIP アドレスに設定。
    • Port:SparkSQL インスタンスへの接続用のポートに設定。
    • TransportMode:SparkSQL サーバーとの通信に使用するトランスポートモード。有効な入力値は、BINARY およびHTTP です。デフォルトではBINARY が選択されます。
    • AuthScheme:使用される認証スキーム。有効な入力値はPLAIN、LDAP、NOSASL、およびKERBEROS です。デフォルトではPLAIN が選択されます。

    Databricks への接続

    Databricks クラスターに接続するには、以下の説明に従ってプロパティを設定します。Note:The needed values can be found in your Databricks instance by navigating to 'Clusters', selecting the desired cluster, and selecting the JDBC/ODBC tab under 'Advanced Options'.

    • Server:Set to the Server Hostname of your Databricks cluster.
    • Port:443
    • TransportMode:HTTP
    • HTTPPath:Set to the HTTP Path of your Databricks cluster.
    • UseSSL:True
    • AuthScheme:PLAIN
    • User:'token' に設定。
    • Password:Set to your personal access token (value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).

Configure the Spark Destination

In the destination component Connection Manager, define mappings from the SQL Server source table into the Spark destination table.

  1. Double-click the Spark destination to open the destination component editor.
  2. In the Connection Managers tab, select the connection manager previously created.
  • Specify the table into which the data will be exported: On the Component Properties tab, enter the table name in the TableName property, under the Custom Properties section. In this case, you will transfer Customers records.
  • On the Column Mappings tab, configure the mappings from the input columns to the destination columns.
  • Run the Project

    You can now run the project. After the SSIS Task has finished executing, data from your SQL table will be exported to the chosen table.

     
     
    ダウンロード