JDBC Mysql Source Connector

Description

Read external data source data through JDBC.

Support Mysql Version

  • 5.5/5.6/5.7/8.0

Support Those Engines

Spark
Flink
SeaTunnel Zeta

Using Dependency

  1. You need to ensure that the jdbc driver jar package has been placed in directory ${SEATUNNEL_HOME}/plugins/.

For SeaTunnel Zeta Engine

  1. You need to ensure that the jdbc driver jar package has been placed in directory ${SEATUNNEL_HOME}/lib/.

Key Features

supports query SQL and can achieve projection effect.

Supported DataSource Info

Datasource Supported versions Driver Url Maven
Mysql Different dependency version has different driver class. com.mysql.cj.jdbc.Driver jdbc:mysql://localhost:3306:3306/test Download

Database Dependency

Please download the support list corresponding to ‘Maven’ and copy it to the ‘$SEATNUNNEL_HOME/plugins/jdbc/lib/‘ working directory
For example Mysql datasource: cp mysql-connector-java-xxx.jar $SEATNUNNEL_HOME/plugins/jdbc/lib/

Data Type Mapping

Mysql Data Type SeaTunnel Data Type
BIT(1)
TINYINT(1)
BOOLEAN
TINYINT BYTE
TINYINT UNSIGNED
SMALLINT
SMALLINT
SMALLINT UNSIGNED
MEDIUMINT
MEDIUMINT UNSIGNED
INT
INTEGER
YEAR
INT
INT UNSIGNED
INTEGER UNSIGNED
BIGINT
BIGINT
BIGINT UNSIGNED DECIMAL(20,0)
DECIMAL(x,y)(Get the designated column’s specified column size.<38) DECIMAL(x,y)
DECIMAL(x,y)(Get the designated column’s specified column size.>38) DECIMAL(38,18)
DECIMAL UNSIGNED DECIMAL((Get the designated column’s specified column size)+1,
(Gets the designated column’s number of digits to right of the decimal point.)))
FLOAT
FLOAT UNSIGNED
FLOAT
DOUBLE
DOUBLE UNSIGNED
DOUBLE
CHAR
VARCHAR
TINYTEXT
MEDIUMTEXT
TEXT
LONGTEXT
JSON
ENUM
STRING
DATE DATE
TIME(s) TIME(s)
DATETIME
TIMESTAMP(s)
TIMESTAMP(s)
TINYBLOB
MEDIUMBLOB
BLOB
LONGBLOB
BINARY
VARBINAR
BIT(n)
GEOMETRY
BYTES

Source Options

Name Type Required Default Description
url String Yes - The URL of the JDBC connection. Refer to a case: jdbc:mysql://localhost:3306:3306/test
driver String Yes - The jdbc class name used to connect to the remote data source,
if you use MySQL the value is com.mysql.cj.jdbc.Driver.
user String No - Connection instance user name
password String No - Connection instance password
query String Yes - Query statement
connection_check_timeout_sec Int No 30 The time in seconds to wait for the database operation used to validate the connection to complete
partition_column String No - The column name for parallelism’s partition, only support numeric type,Only support numeric type primary key, and only can config one column.
partition_lower_bound BigDecimal No - The partition_column min value for scan, if not set SeaTunnel will query database get min value.
partition_upper_bound BigDecimal No - The partition_column max value for scan, if not set SeaTunnel will query database get max value.
partition_num Int No job parallelism The number of partition count, only support positive integer. default value is job parallelism
fetch_size Int No 0 For queries that return a large number of objects,you can configure
the row fetch size used in the query toimprove performance by
reducing the number database hits required to satisfy the selection criteria.
Zero means use jdbc default value.
properties Map No - Additional connection configuration parameters,when properties and URL have the same parameters, the priority is determined by the
specific implementation of the driver. For example, in MySQL, properties take precedence over the URL.
table_path Int No 0 The path to the full path of table, you can use this configuration instead of query.
examples:
mysql: “testdb.table1”
oracle: “test_schema.table1”
sqlserver: “testdb.test_schema.table1”
postgresql: “testdb.test_schema.table1”
table_list Array No 0 The list of tables to be read, you can use this configuration instead of table_path example: [{ table_path = "testdb.table1"}, {table_path = "testdb.table2", query = "select * id, name from testdb.table2"}]
where_condition String No - Common row filter conditions for all tables/queries, must start with where. for example where id > 100
split.size Int No 8096 The split size (number of rows) of table, captured tables are split into multiple splits when read of table.
split.even-distribution.factor.lower-bound Double No 0.05 The lower bound of the chunk key distribution factor. This factor is used to determine whether the table data is evenly distributed. If the distribution factor is calculated to be greater than or equal to this lower bound (i.e., (MAX(id) - MIN(id) + 1) / row count), the table chunks would be optimized for even distribution. Otherwise, if the distribution factor is less, the table will be considered as unevenly distributed and the sampling-based sharding strategy will be used if the estimated shard count exceeds the value specified by sample-sharding.threshold. The default value is 0.05.
split.even-distribution.factor.upper-bound Double No 100 The upper bound of the chunk key distribution factor. This factor is used to determine whether the table data is evenly distributed. If the distribution factor is calculated to be less than or equal to this upper bound (i.e., (MAX(id) - MIN(id) + 1) / row count), the table chunks would be optimized for even distribution. Otherwise, if the distribution factor is greater, the table will be considered as unevenly distributed and the sampling-based sharding strategy will be used if the estimated shard count exceeds the value specified by sample-sharding.threshold. The default value is 100.0.
split.sample-sharding.threshold Int No 10000 This configuration specifies the threshold of estimated shard count to trigger the sample sharding strategy. When the distribution factor is outside the bounds specified by chunk-key.even-distribution.factor.upper-bound and chunk-key.even-distribution.factor.lower-bound, and the estimated shard count (calculated as approximate row count / chunk size) exceeds this threshold, the sample sharding strategy will be used. This can help to handle large datasets more efficiently. The default value is 1000 shards.
split.inverse-sampling.rate Int No 1000 The inverse of the sampling rate used in the sample sharding strategy. For example, if this value is set to 1000, it means a 1/1000 sampling rate is applied during the sampling process. This option provides flexibility in controlling the granularity of the sampling, thus affecting the final number of shards. It’s especially useful when dealing with very large datasets where a lower sampling rate is preferred. The default value is 1000.
common-options No - Source plugin common parameters, please refer to Source Common Options for details

Parallel Reader

The JDBC Source connector supports parallel reading of data from tables. SeaTunnel will use certain rules to split the data in the table, which will be handed over to readers for reading. The number of readers is determined by the parallelism option.

Split Key Rules:

  1. If partition_column is not null, It will be used to calculate split. The column must in Supported split data type.
  2. If partition_column is null, seatunnel will read the schema from table and get the Primary Key and Unique Index. If there are more than one column in Primary Key and Unique Index, The first column which in the supported split data type will be used to split data. For example, the table have Primary Key(nn guid, name varchar), because guid id not in supported split data type, so the column name will be used to split data.

Supported split data type:

  • String
  • Number(int, bigint, decimal, …)
  • Date

split.size

How many rows in one split, captured tables are split into multiple splits when read of table.

split.even-distribution.factor.lower-bound

Not recommended for use

The lower bound of the chunk key distribution factor. This factor is used to determine whether the table data is evenly distributed. If the distribution factor is calculated to be greater than or equal to this lower bound (i.e., (MAX(id) - MIN(id) + 1) / row count), the table chunks would be optimized for even distribution. Otherwise, if the distribution factor is less, the table will be considered as unevenly distributed and the sampling-based sharding strategy will be used if the estimated shard count exceeds the value specified by sample-sharding.threshold. The default value is 0.05.

split.even-distribution.factor.upper-bound

Not recommended for use

The upper bound of the chunk key distribution factor. This factor is used to determine whether the table data is evenly distributed. If the distribution factor is calculated to be less than or equal to this upper bound (i.e., (MAX(id) - MIN(id) + 1) / row count), the table chunks would be optimized for even distribution. Otherwise, if the distribution factor is greater, the table will be considered as unevenly distributed and the sampling-based sharding strategy will be used if the estimated shard count exceeds the value specified by sample-sharding.threshold. The default value is 100.0.

split.sample-sharding.threshold

This configuration specifies the threshold of estimated shard count to trigger the sample sharding strategy. When the distribution factor is outside the bounds specified by chunk-key.even-distribution.factor.upper-bound and chunk-key.even-distribution.factor.lower-bound, and the estimated shard count (calculated as approximate row count / chunk size) exceeds this threshold, the sample sharding strategy will be used. This can help to handle large datasets more efficiently. The default value is 1000 shards.

split.inverse-sampling.rate

The inverse of the sampling rate used in the sample sharding strategy. For example, if this value is set to 1000, it means a 1/1000 sampling rate is applied during the sampling process. This option provides flexibility in controlling the granularity of the sampling, thus affecting the final number of shards. It’s especially useful when dealing with very large datasets where a lower sampling rate is preferred. The default value is 1000.

partition_column [string]

The column name for split data.

partition_upper_bound [BigDecimal]

The partition_column max value for scan, if not set SeaTunnel will query database get max value.

partition_lower_bound [BigDecimal]

The partition_column min value for scan, if not set SeaTunnel will query database get min value.

partition_num [int]

Not recommended for use, The correct approach is to control the number of split through split.size

How many splits do we need to split into, only support positive integer. default value is job parallelism.

tips

If the table can not be split(for example, table have no Primary Key or Unique Index, and partition_column is not set), it will run in single concurrency.

Use table_path to replace query for single table reading. If you need to read multiple tables, use table_list.

Task Example

Simple:

This example queries type_bin ‘table’ 16 data in your test “database” in single parallel and queries all of its fields. You can also specify which fields to query for final output to the console.

  1. # Defining the runtime environment
  2. env {
  3. parallelism = 4
  4. job.mode = "BATCH"
  5. }
  6. source{
  7. Jdbc {
  8. url = "jdbc:mysql://localhost:3306/test?serverTimezone=GMT%2b8&useUnicode=true&characterEncoding=UTF-8&rewriteBatchedStatements=true"
  9. driver = "com.mysql.cj.jdbc.Driver"
  10. connection_check_timeout_sec = 100
  11. user = "root"
  12. password = "123456"
  13. query = "select * from type_bin limit 16"
  14. }
  15. }
  16. transform {
  17. # If you would like to get more information about how to configure seatunnel and see full list of transform plugins,
  18. # please go to https://seatunnel.apache.org/docs/transform-v2/sql
  19. }
  20. sink {
  21. Console {}
  22. }

parallel by partition_column

  1. env {
  2. parallelism = 4
  3. job.mode = "BATCH"
  4. }
  5. source {
  6. Jdbc {
  7. url = "jdbc:mysql://localhost/test?serverTimezone=GMT%2b8"
  8. driver = "com.mysql.cj.jdbc.Driver"
  9. connection_check_timeout_sec = 100
  10. user = "root"
  11. password = "123456"
  12. query = "select * from type_bin"
  13. partition_column = "id"
  14. split.size = 10000
  15. # Read start boundary
  16. #partition_lower_bound = ...
  17. # Read end boundary
  18. #partition_upper_bound = ...
  19. }
  20. }
  21. sink {
  22. Console {}
  23. }

parallel by Primary Key or Unique Index

Configuring table_path will turn on auto split, you can configure split.* to adjust the split strategy

  1. env {
  2. parallelism = 4
  3. job.mode = "BATCH"
  4. }
  5. source {
  6. Jdbc {
  7. url = "jdbc:mysql://localhost/test?serverTimezone=GMT%2b8"
  8. driver = "com.mysql.cj.jdbc.Driver"
  9. connection_check_timeout_sec = 100
  10. user = "root"
  11. password = "123456"
  12. table_path = "testdb.table1"
  13. query = "select * from testdb.table1"
  14. split.size = 10000
  15. }
  16. }
  17. sink {
  18. Console {}
  19. }

Parallel Boundary:

It is more efficient to specify the data within the upper and lower bounds of the query It is more efficient to read your data source according to the upper and lower boundaries you configured

  1. source {
  2. Jdbc {
  3. url = "jdbc:mysql://localhost:3306/test?serverTimezone=GMT%2b8&useUnicode=true&characterEncoding=UTF-8&rewriteBatchedStatements=true"
  4. driver = "com.mysql.cj.jdbc.Driver"
  5. connection_check_timeout_sec = 100
  6. user = "root"
  7. password = "123456"
  8. # Define query logic as required
  9. query = "select * from type_bin"
  10. partition_column = "id"
  11. # Read start boundary
  12. partition_lower_bound = 1
  13. # Read end boundary
  14. partition_upper_bound = 500
  15. partition_num = 10
  16. properties {
  17. useSSL=false
  18. }
  19. }
  20. }

Multiple table read:

Configuring table_list will turn on auto split, you can configure `split.` to adjust the split strategy*

  1. env {
  2. job.mode = "BATCH"
  3. parallelism = 4
  4. }
  5. source {
  6. Jdbc {
  7. url = "jdbc:mysql://localhost/test?serverTimezone=GMT%2b8"
  8. driver = "com.mysql.cj.jdbc.Driver"
  9. connection_check_timeout_sec = 100
  10. user = "root"
  11. password = "123456"
  12. table_list = [
  13. {
  14. table_path = "testdb.table1"
  15. },
  16. {
  17. table_path = "testdb.table2"
  18. # Use query filetr rows & columns
  19. query = "select id, name from testdb.table2 where id > 100"
  20. }
  21. ]
  22. #where_condition= "where id > 100"
  23. #split.size = 8096
  24. #split.even-distribution.factor.upper-bound = 100
  25. #split.even-distribution.factor.lower-bound = 0.05
  26. #split.sample-sharding.threshold = 1000
  27. #split.inverse-sampling.rate = 1000
  28. }
  29. }
  30. sink {
  31. Console {}
  32. }