Socket source connector
Support Those Engines
Spark
Flink
SeaTunnel Zeta
Key features
Description
Used to read data from Socket.
Data Type Mapping
The File does not have a specific type list, and we can indicate which SeaTunnel data type the corresponding data needs to be converted to by specifying the Schema in the config.
| SeaTunnel Data type | 
|---|
| STRING | 
| SHORT | 
| INT | 
| BIGINT | 
| BOOLEAN | 
| DOUBLE | 
| DECIMAL | 
| FLOAT | 
| DATE | 
| TIME | 
| TIMESTAMP | 
| BYTES | 
| ARRAY | 
| MAP | 
Options
| Name | Type | Required | Default | Description | 
|---|---|---|---|---|
| host | String | Yes | _ | socket server host | 
| port | Integer | Yes | _ | socket server port | 
| common-options | no | - | Source plugin common parameters, please refer to Source Common Options for details. | 
How to Create a Socket Data Synchronization Jobs
- Configuring the SeaTunnel config file
 
The following example demonstrates how to create a data synchronization job that reads data from Socket and prints it on the local client:
# Set the basic configuration of the task to be performedenv {parallelism = 1job.mode = "BATCH"}# Create a source to connect to socketsource {Socket {host = "localhost"port = 9999}}# Console printing of the read socket datasink {Console {parallelism = 1}}
- Start a port listening
 
nc -l 9999
Start a SeaTunnel task
Socket Source send test data
~ nc -l 9999testhelloflinkspark
- Console Sink print data
 
[test][hello][flink][spark]
