AWS Glue Jobs からSage US Data にJDBC 経由でデータ連携

Amazon S3 でホストされているCData JDBC ドライバーを使用してAWS Glue ジョブからSage US にデータ連携。

AWS Glue はAmazon のETL サービスであり、これを使用すると、簡単にデータプレパレーションを行い、ストレージおよび分析用に読み込むことができます。AWS Glue と一緒にPySpark モジュールを使用すると、JDBC 接続を経由でデータを処理するジョブを作成し、そのデータをAWS データストアに直接読み込むことができます。ここでは、CData JDBC Driver for Sage US をAmazon S3 バケットにアップロードし、Sage US data からデータを抽出してCSV ファイルとしてS3 に保存するためのAWS Glue ジョブを作成して実行する方法について説明します。

CData JDBC driver for Sage US をAmazon S3 バケットにアップロード

In order to work with the CData JDBC Driver for Sage US in AWS Glue, you will need to store it (and any relevant license files) in a bucket in Amazon S3.

  1. Open the Amazon S3 Console.
  2. Select an existing bucket (or create a new one).
  3. Click Upload
  4. Select the JAR file (cdata.jdbc.sage50us.jar) found in the lib directory in the installation location for the driver.

Amazon Glue Job を設定

  1. Navigate to ETL -> Jobs from the AWS Glue Console.
  2. Click Add Job to create a new Glue job.
  3. Fill in the Job properties:
    • Name: Fill in a name for the job, for example: Sage50USGlueJob.
    • IAM Role: Select (or create) an IAM role that has the AWSGlueServiceRole and AmazonS3FullAccess (because the JDBC Driver and destination are in an Amazon S3 bucket) permissions policies.
    • Type: Select "Spark."
    • This job runs: Select "A new script to be authored by you".
      Populate the script properties:
      • Script file name: A name for the script file, for example:GlueSage50USJDBC
      • S3 path where the script is stored: Fill in or browse to an S3 bucket.
      • Temporary directory: Fill in or browse to an S3 bucket.
    • ETL language: Select "Python."
    • Expand Security configuration, script libraries and job parameters (optional).For Dependent jars path, fill in or browse to the S3 bucket where you loaded the JAR file.Be sure to include the name of the JAR file itself in the path, i.e.: s3://mybucket/cdata.jdbc.sage50us.jar
  4. Click Next.Here you will have the option to add connection to other AWS endpoints, so if your Destination is Redshift, MySQL, etc, you can create and use connections to those data sources.
  5. Click "Save job and edit script" to create the job.
  6. In the editor that opens, write a python script for the job.You can use the sample script (see below) as an example.

サンプルGlue スクリプト

To connect to Sage US using the CData JDBC driver, you will need to create a JDBC URL, populating the necessary connection properties.Additionally, (unless you are using a Beta driver), you will need to set the RTK property in the JDBC URL.You can view the licensing file included in the installation for information on how to set this property.

The Application Id and Company Name connection string options are required to connect to Sage as a data source. You can obtain an Application Id by contacting Sage directly to request access to the Sage 50 SDK.

Sage must be installed on the machine. The Sage.Peachtree.API.dll and Sage.Peachtree.API.Resolver.dll assemblies are required. These assemblies are installed with Sage in C:\Program Files\Sage\Peachtree\API\. Additionally, the Sage SDK requires .NET Framework 4.0 and is only compatible with 32-bit applications. To use the Sage SDK in Visual Studio, set the Platform Target property to "x86" in Project -> Properties -> Build.

You must authorize the application to access company data: To authorize your application to access Sage, restart the Sage application, open the company you want to access, and connect with your application. You will then be prompted to set access permissions for the application in the resulting dialog.

While the compiled executable will require authorization only once, during development you may need to follow this process to reauthorize a new build. To avoid restarting the Sage application when developing with Visual Studio, click Build -> Configuration Manager and uncheck "Build" for your project.

ビルトイン接続文字列デザイナー

For assistance in constructing the JDBC URL, use the connection string designer built into the Sage US JDBC Driver.Either double-click the JAR file or execute the jar file from the command-line.

java -jar cdata.jdbc.sage50us.jar

Fill in the connection properties and copy the connection string to the clipboard.

To host the JDBC driver in Amazon S3, you will need a license (full or trial) and a Runtime Key (RTK).For more information on obtaining this license (or a trial), contact our sales team.

Below is a sample script that uses the CData JDBC driver with the PySpark and AWSGlue modules to extract Sage US data and write it to an S3 bucket in CSV format.Make any changes to the script you need to suit your needs and save the job.

import sys from awsglue.transforms import * from awsglue.utils import getResolvedOptions from pyspark.context import SparkContext from awsglue.context import GlueContext from awsglue.dynamicframe import DynamicFrame from awsglue.job import Job args = getResolvedOptions(sys.argv, ['JOB_NAME']) sparkContext = SparkContext() glueContext = GlueContext(sparkContext) sparkSession = glueContext.spark_session ##Use the CData JDBC driver to read Sage US data from the Customer table into a DataFrame ##Note the populated JDBC URL and driver class name source_df = sparkSession.read.format("jdbc").option("url","jdbc:sage50us:RTK=5246...;ApplicationId=8dfafu4V4ODmh1fM0xx;CompanyName=Bellwether Garden Supply - Premium;").option("dbtable","Customer").option("driver","cdata.jdbc.sage50us.Sage50USDriver").load() glueJob = Job(glueContext) glueJob.init(args['JOB_NAME'], args) ##Convert DataFrames to AWS Glue's DynamicFrames Object dynamic_dframe = DynamicFrame.fromDF(source_df, glueContext, "dynamic_df") ##Write the DynamicFrame as a file in CSV format to a folder in an S3 bucket. ##It is possible to write to any Amazon data store (SQL Server, Redshift, etc) by using any previously defined connections. retDatasink4 = glueContext.write_dynamic_frame.from_options(frame = dynamic_dframe, connection_type = "s3", connection_options = {"path": "s3://mybucket/outfiles"}, format = "csv", transformation_ctx = "datasink4") glueJob.commit()

Glueジョブを実行する

With the script written, we are ready to run the Glue job.Click Run Job and wait for the extract/load to complete.You can view the status of the job from the Jobs page in the AWS Glue Console.Once the Job has succeeded, you will have a csv file in your S3 bucket with data from the Sage US Customer table.

Using the CData JDBC Driver for Sage US in AWS Glue, you can easily create ETL jobs for Sage US data, writing the data to an S3 bucket or loading it into any other AWS data store.

 
 
ダウンロード