Devlive 开源社区 本次搜索耗时 0.563 秒,为您找到 282 个相关结果.
  • External Link Management

    Background How to create the external link Where can see the external link Background In production practice, in order to manage the Flink job properly, there is always a need...
  • Branching and Tagging

    3981 2025-03-10 《Apache Iceberg 1.8.1》
    Branching and Tagging Overview Iceberg table metadata maintains a snapshot log, which represents the changes applied to a table. Snapshots are fundamental in Iceberg as they are ...
  • Event Listener

    Introduction Support Those Engines API Event Data API Event Listener API Event Collect API Configuration Listener Zeta Engine Flink Engine Spark Engine Introduction The...
  • Dinky

    Dinky Overview Create Task Task Parameter Task Example Dinky Task Example Dinky Overview Dinky 任务类型,用于创建并执行Dinky 类型任务以支撑一站式的开发、调试、运维 FlinkSQL、Flink Jar、SQL。worker 执行该任务的时候...
  • Seatunnel runs on Flink

    Set Flink configuration information in the job How to set up a simple Flink job How to run a job in a project Flink is a powerful high-performance distributed stream processing...
  • Building A Runnable Distribution

    3659 2024-07-05 《Apache Kyuubi 1.9.1》
    To create a Kyuubi distribution like those distributed by Kyuubi Release Page , and that is laid out to be runnable, use ./build/dist in the project root directory. For more inf...
  • Flink Configuration

    3659 2025-03-12 《Apache Iceberg 1.8.1》
    Flink Configuration Catalog Configuration A catalog is created and named by executing the following query (replace <catalog_name> with your catalog name and <config_key> =<confi...
  • Slack

    Support Those Engines Key features Description Data Type Mapping Options Task Example Simple: Changelog new version Slack sink connector Support Those Engines Spark...
  • AmazonSqs

    Support Those Engines Description Key Features Sink Options Task Example Amazon SQS sink connector Support Those Engines Spark Flink SeaTunnel Zeta Description Writ...
  • DDL

    3571 2025-03-11 《Apache Iceberg 1.8.1》
    Spark DDL To use Iceberg in Spark, first configure Spark catalogs . Iceberg uses Apache Spark’s DataSourceV2 API for data source and catalog implementations. CREATE TABLE Spark...