Devlive 开源社区 本次搜索耗时 0.851 秒,为您找到 148 个相关结果.
  • Redis Connector

    Redis Write Dependency Writing Redis the Regular Way 1.Access to source 2. Write to redis Apache StreamPark™ Writes to Redis 1. Configure policy and connection information 2. ...
  • Introduction

    6076 2024-06-24 《Apache Amoro 0.6.1》
    Architecture Supported table formats Supported engines Iceberg format Paimon format Mixed format User cases Self-managed streaming Lakehouse Stream-and-batch-fused data pipe...
  • 编程模型

    架构 编程模型 DataStream Flink Sql TableEnvironment StreamTableEnvironment RunTime Context StreamingContext TableContext StreamTableContext 生命周期 生命周期之 — init 生命周期之 — config 生...
  • Concurrency Control

    5878 2024-06-28 《Apache Hudi 0.15.0》
    Deployment models with supported concurrency controls Model A: Single writer with inline table services Single Writer Guarantees Model B: Single writer with async table services ...
  • SQL Server CDC

    Support SQL Server Version Support Those Engines Key Features Description Supported DataSource Info Using Dependency Install Jdbc Driver For Spark/Flink Engine For SeaTunnel ...
  • Set Up with Kubernetes

    Prerequisites Installation SeaTunnel docker image Flink Zeta (local-mode) Zeta (cluster-mode) Deploying the operator Flink Zeta (local-mode) Zeta (cluster-mode) Run SeaTu...
  • S3File

    Support Those Engines Key Features Description Supported DataSource Info Database Dependency Data Type Mapping Orc File Type Parquet File Type Sink Options path [string] h...
  • General

    5673 2024-07-01 《Apache Hudi 0.15.0》
    When is Hudi useful for me or my organization? What are some non-goals for Hudi? What is incremental processing? Why does Hudi docs/talks keep talking about it? How is Hudi opti...
  • JDBC Connector

    JDBC 信息配置 semantic 语义配置 EXACTLY_ONCE AT_LEAST_ONCE && NONE 其他配置 JDBC 读取数据 queryFunc获取一条sql resultFunc 处理查询到的数据 JDBC 读取写入 根据数据流生成目标SQL 设置写入批次大小 多实例 JDBC 支持 手动指定 JDBC 连接信...
  • Troubleshooting

    5517 2024-07-01 《Apache Hudi 0.15.0》
    Writing Tables org.apache.parquet.io.InvalidRecordException: Parquet/Avro schema mismatch: Avro field ‘col1’ not found java.lang.UnsupportedOperationException: org.apache.parquet....