mirror of
https://gitee.com/devlive-community/datacap.git
synced 2024-12-14 17:30:51 +08:00
165 lines
5.1 KiB
YAML
165 lines
5.1 KiB
YAML
name: ClickHouse
|
|
supportTime: '2022-09-22'
|
|
|
|
configures:
|
|
- field: name
|
|
type: String
|
|
required: true
|
|
message: name is a required field, please be sure to enter
|
|
- field: host
|
|
type: String
|
|
required: true
|
|
value: 127.0.0.1
|
|
message: host is a required field, please be sure to enter
|
|
- field: port
|
|
type: Number
|
|
required: true
|
|
min: 1
|
|
max: 65535
|
|
value: 9000
|
|
message: port is a required field, please be sure to enter
|
|
- field: username
|
|
type: String
|
|
group: authorization
|
|
- field: password
|
|
type: String
|
|
group: authorization
|
|
- field: database
|
|
type: String
|
|
group: advanced
|
|
value: default
|
|
required: true
|
|
- field: configures
|
|
type: Array
|
|
value: [ ]
|
|
group: custom
|
|
|
|
pipelines:
|
|
- executor: Seatunnel
|
|
type: SOURCE
|
|
fields:
|
|
- field: host
|
|
origin: host|port
|
|
required: true
|
|
- field: username
|
|
origin: username
|
|
required: true
|
|
override: false
|
|
- field: password
|
|
origin: password
|
|
required: false
|
|
override: false
|
|
- field: database
|
|
origin: database
|
|
required: true
|
|
hidden: false
|
|
description: The ClickHouse database
|
|
override: true
|
|
input: true
|
|
width: 300
|
|
type: INPUT
|
|
- field: sql
|
|
origin: context
|
|
required: false
|
|
hidden: false
|
|
description: The query sql used to search data though Clickhouse server
|
|
override: true
|
|
input: true
|
|
width: 300
|
|
type: TEXT
|
|
tooltip: If this value is entered, the default sql statement will be replaced
|
|
- field: result_table_name
|
|
origin: result_table_name
|
|
description: |
|
|
When result_table_name is not specified, the data processed by this plugin will not be registered as a data set (dataStream/dataset) that can be directly accessed by other plugins, or called a temporary table (table) ;
|
|
When result_table_name is specified, the data processed by this plugin will be registered as a data set (dataStream/dataset) that can be directly accessed by other plugins, or called a temporary table (table) . The data set (dataStream/dataset) registered here can be directly accessed by other plugins by specifying source_table_name .
|
|
required: false
|
|
hidden: true
|
|
input: true
|
|
override: true
|
|
- executor: Seatunnel
|
|
type: SINK
|
|
fields:
|
|
- field: host
|
|
origin: host|port
|
|
required: true
|
|
- field: username
|
|
origin: username
|
|
override: false
|
|
required: true
|
|
- field: password
|
|
override: false
|
|
required: true
|
|
- field: database
|
|
origin: database
|
|
description: |
|
|
The ClickHouse database
|
|
override: true
|
|
required: true
|
|
input: true
|
|
- field: table
|
|
origin: table
|
|
description: |
|
|
The table name of the ClickHouse database
|
|
override: true
|
|
required: true
|
|
input: true
|
|
- field: fields
|
|
origin: fields
|
|
description: |
|
|
The data field that needs to be output to ClickHouse , if not configured, it will be automatically adapted according to the sink table schema .
|
|
|
|
If there are multiple values, each value needs to be one row.
|
|
override: true
|
|
required: true
|
|
input: true
|
|
type: TEXT
|
|
- field: bulk_size
|
|
origin: bulk_size
|
|
description: |
|
|
The number of rows written through Clickhouse-jdbc each time, the default is 20000 .
|
|
override: true
|
|
required: false
|
|
input: true
|
|
value: 20000
|
|
width: 100
|
|
type: NUMBER
|
|
- field: split_mode
|
|
origin: split_mode
|
|
description: |
|
|
This mode only support clickhouse table which engine is 'Distributed'.And internal_replication option should be true. They will split distributed table data in seatunnel and perform write directly on each shard. The shard weight define is clickhouse will be counted.
|
|
override: true
|
|
required: false
|
|
input: true
|
|
value: false
|
|
width: 45
|
|
type: SWITCH
|
|
- field: sharding_key
|
|
origin: sharding_key
|
|
description: |
|
|
When use split_mode, which node to send data to is a problem, the default is random selection, but the 'sharding_key' parameter can be used to specify the field for the sharding algorithm. This option only worked when 'split_mode' is true.
|
|
override: true
|
|
required: false
|
|
input: true
|
|
width: 100
|
|
type: INPUT
|
|
- field: primary_key
|
|
origin: primary_key
|
|
description: |
|
|
Mark the primary key column from clickhouse table, and based on primary key execute INSERT/UPDATE/DELETE to clickhouse table
|
|
override: true
|
|
required: false
|
|
input: true
|
|
width: 100
|
|
type: INPUT
|
|
- field: support_upsert
|
|
origin: support_upsert
|
|
description: |
|
|
Support upsert row by query primary key
|
|
override: true
|
|
required: false
|
|
input: true
|
|
value: false
|
|
width: 45
|
|
type: SWITCH
|