Devlive 开源社区 本次搜索耗时 0.058 秒,为您找到 2342 个相关结果.
  • SQL

    SQL 综述 创建数据源 创建任务 任务参数 任务样例 Hive表创建示例 在hive中创建临时表并写入数据 运行该任务成功之后在hive中查询结果 使用前置sql和后置sql示例 注意事项 SQL 综述 SQL任务类型,用于连接数据库并执行相应SQL。 创建数据源 可参考 数据源配置 数据源中心 。 创建任务 点击项...
  • PostgreSQL

    Support Those Engines Using Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key Features Description Supported DataSource Info Database Dependency Data Type Map...
  • Setup

    6316 2024-06-21 《Apache Accumulo 2.x》
    User Manual (2.x and 3.x) Master/Manager naming Setup for testing or development Setup for Production Configuring Accumulo Initialization Run Accumulo Run individual Accumulo...
  • Design & Concepts

    6293 2024-07-01 《Apache Hudi 0.15.0》
    How does Hudi ensure atomicity? Does Hudi extend the Hive table layout? What concurrency control approaches does Hudi adopt? Hudi’s commits are based on transaction start time i...
  • Troubleshooting

    6282 2024-07-01 《Apache Hudi 0.15.0》
    Writing Tables org.apache.parquet.io.InvalidRecordException: Parquet/Avro schema mismatch: Avro field ‘col1’ not found java.lang.UnsupportedOperationException: org.apache.parquet....
  • Flink Getting Started

    6260 2024-06-29 《Apache Iceberg 1.5.2》
    Preparation when using Flink SQL Client Flink’s Python API Adding catalogs. Catalog Configuration Hive catalog Creating a table Writing Branch Writes Reading Type conversi...
  • Clustering

    6258 2024-06-30 《Apache Hudi 0.15.0》
    Background How is compaction different from clustering? Clustering Architecture Overall, there are 2 steps to clustering Schedule clustering Execute clustering Clustering Use...
  • SQL DDL

    6249 2024-06-30 《Apache Hudi 0.15.0》
    Spark SQL Create table Create non-partitioned table Create partitioned table Create table with record keys and ordering fields Create table from an external location Create Ta...
  • Docker 部署

    前置条件 1. 安装 docker 2. 安装 docker-compose 部署 Apache StreamPark™ 1. 基于 h2 和 docker-compose 部署 Apache StreamPark™ 2. 部署 3. 配置flink home 4. 配置session集群 5. 提交 Flink 作业 使用已有的 Mysql ...
  • Oracle

    Description Support Those Engines Using Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key Features Supported DataSource Info Database Dependency Data Type Map...