跳到主要内容
版本:2.3.3

Oracle

JDBC Oracle Sink Connector

Support Those Engines

Spark
Flink
SeaTunnel Zeta

Key Features

Use Xa transactions to ensure exactly-once. So only support exactly-once for the database which is support Xa transactions. You can set is_exactly_once=true to enable it.

Description

Write data through jdbc. Support Batch mode and Streaming mode, support concurrent writing, support exactly-once semantics (using XA transaction guarantee).

Supported DataSource Info

DatasourceSupported VersionsDriverUrlMaven
OracleDifferent dependency version has different driver class.oracle.jdbc.OracleDriverjdbc:oracle:thin:@datasource01:1523:xehttps://mvnrepository.com/artifact/com.oracle.database.jdbc/ojdbc8

Database Dependency

Please download the support list corresponding to 'Maven' and copy it to the '$SEATNUNNEL_HOME/plugins/jdbc/lib/' working directory
For example Oracle datasource: cp ojdbc8-xxxxxx.jar $SEATNUNNEL_HOME/lib/
To support the i18n character set, copy the orai18n.jar to the $SEATNUNNEL_HOME/lib/ directory.

Data Type Mapping

PostgreSQL Data typeSeaTunnel Data type
INTEGERINT
FLOATDECIMAL(38, 18)
NUMBER(precision <= 9, scale == 0)INT
NUMBER(9 < precision <= 18, scale == 0)BIGINT
NUMBER(18 < precision, scale == 0)DECIMAL(38, 0)
NUMBER(scale != 0)DECIMAL(38, 18)
BINARY_DOUBLEDOUBLE
BINARY_FLOAT
REAL
FLOAT
CHAR
NCHAR
NVARCHAR2
VARCHAR2
LONG
ROWID
NCLOB
CLOB
STRING
DATEDATE
TIMESTAMP
TIMESTAMP WITH LOCAL TIME ZONE
TIMESTAMP
BLOB
RAW
LONG RAW
BFILE
BYTES

Options

NameTypeRequiredDefaultDescription
urlStringYes-The URL of the JDBC connection. Refer to a case: jdbc:postgresql://localhost:5432/test
driverStringYes-The jdbc class name used to connect to the remote data source,
if you use Oracle the value is oracle.jdbc.OracleDriver.
userStringNo-Connection instance user name
passwordStringNo-Connection instance password
queryStringNo-Use this sql write upstream input datas to database. e.g INSERT ...,query have the higher priority
databaseStringNo-Use this database and table-name auto-generate sql and receive upstream input datas write to database.
This option is mutually exclusive with query and has a higher priority.
tableStringNo-Use database and this table-name auto-generate sql and receive upstream input datas write to database.
This option is mutually exclusive with query and has a higher priority.
primary_keysArrayNo-This option is used to support operations such as insert, delete, and update when automatically generate sql.
support_upsert_by_query_primary_key_existBooleanNofalseChoose to use INSERT sql, UPDATE sql to process update events(INSERT, UPDATE_AFTER) based on query primary key exists. This configuration is only used when database unsupport upsert syntax. Note: that this method has low performance
connection_check_timeout_secIntNo30The time in seconds to wait for the database operation used to validate the connection to complete.
max_retriesIntNo0The number of retries to submit failed (executeBatch)
batch_sizeIntNo1000For batch writing, when the number of buffered records reaches the number of batch_size or the time reaches batch_interval_ms
, the data will be flushed into the database
batch_interval_msIntNo1000For batch writing, when the number of buffers reaches the number of batch_size or the time reaches batch_interval_ms, the data will be flushed into the database
is_exactly_onceBooleanNofalseWhether to enable exactly-once semantics, which will use Xa transactions. If on, you need to
set xa_data_source_class_name.
generate_sink_sqlBooleanNofalseGenerate sql statements based on the database table you want to write to.
xa_data_source_class_nameStringNo-The xa data source class name of the database Driver, for example, Oracle is oracle.jdbc.xa.client.OracleXADataSource, and
please refer to appendix for other data sources
max_commit_attemptsIntNo3The number of retries for transaction commit failures
transaction_timeout_secIntNo-1The timeout after the transaction is opened, the default is -1 (never timeout). Note that setting the timeout may affect
exactly-once semantics
auto_commitBooleanNotrueAutomatic transaction commit is enabled by default
common-optionsno-Sink plugin common parameters, please refer to Sink Common Options for details

Tips

If partition_column is not set, it will run in single concurrency, and if partition_column is set, it will be executed in parallel according to the concurrency of tasks.

Task Example

Simple:

This example defines a SeaTunnel synchronization task that automatically generates data through FakeSource and sends it to JDBC Sink. FakeSource generates a total of 16 rows of data (row.num=16), with each row having two fields, name (string type) and age (int type). The final target table is test_table will also be 16 rows of data in the table. Before run this job, you need create database test and table test_table in your PostgreSQL. And if you have not yet installed and deployed SeaTunnel, you need to follow the instructions in Install SeaTunnel to install and deploy SeaTunnel. And then follow the instructions in Quick Start With SeaTunnel Engine to run this job.

# Defining the runtime environment
env {
# You can set flink configuration here
execution.parallelism = 1
job.mode = "BATCH"
}

source {
FakeSource {
parallelism = 1
result_table_name = "fake"
row.num = 16
schema = {
fields {
name = "string"
age = "int"
}
}
}
# If you would like to get more information about how to configure seatunnel and see full list of source plugins,
# please go to https://seatunnel.apache.org/docs/category/source-v2
}

transform {
# If you would like to get more information about how to configure seatunnel and see full list of transform plugins,
# please go to https://seatunnel.apache.org/docs/category/transform-v2
}

sink {
jdbc {
url = "jdbc:oracle:thin:@datasource01:1523:xe"
driver = "oracle.jdbc.OracleDriver"
user = root
password = 123456
query = "INSERT INTO TEST.TEST_TABLE(NAME,AGE) VALUES(?,?)"
}
# If you would like to get more information about how to configure seatunnel and see full list of sink plugins,
# please go to https://seatunnel.apache.org/docs/category/sink-v2
}

Generate Sink SQL

This example not need to write complex sql statements, you can configure the database name table name to automatically generate add statements for you

sink {
Jdbc {
url = "jdbc:oracle:thin:@datasource01:1523:xe"
driver = "oracle.jdbc.OracleDriver"
user = root
password = 123456

generate_sink_sql = true
database = XE
table = "TEST.TEST_TABLE"
}
}

Exactly-once :

For accurate write scene we guarantee accurate once

sink {
jdbc {
url = "jdbc:oracle:thin:@datasource01:1523:xe"
driver = "oracle.jdbc.OracleDriver"

max_retries = 0
user = root
password = 123456
query = "INSERT INTO TEST.TEST_TABLE(NAME,AGE) VALUES(?,?)"

is_exactly_once = "true"

xa_data_source_class_name = "oracle.jdbc.xa.client.OracleXADataSource"
}
}

CDC(Change Data Capture) Event

CDC change data is also supported by us In this case, you need config database, table and primary_keys.

sink {
jdbc {
url = "jdbc:oracle:thin:@datasource01:1523:xe"
driver = "oracle.jdbc.OracleDriver"
user = root
password = 123456

generate_sink_sql = true
# You need to configure both database and table
database = XE
table = "TEST.TEST_TABLE"
primary_keys = ["ID"]
}
}