Devlive 开源社区 本次搜索耗时 0.360 秒,为您找到 1306 个相关结果.
  • CosFile

    Description Key Features Options path [string] bucket [string] secret_id [string] secret_key [string] region [string] custom_filename [boolean] file_name_expression [string...
  • My Hours

    Support Those Engines Key Features Description Key features Supported DataSource Info Source Options How to Create a My Hours Data Synchronization Jobs Parameter Interpretat...
  • Using Tables

    5237 2024-06-26 《Apache Amoro 0.6.1》
    Create table Configure LogStore Configure watermark Modify table Upgrade a Hive table Configure self-optimizing Modify optimizer group Adjust optimizing resources Adjust opt...
  • Attach-Detach

    5232 2024-05-25 《Apache JDO 3.2.1》
    Detach All On Commit Copy On Attach Serialization of Detachable classes JDO provides an interface to the persistence of objects. JDO 1.0 didn’t provide a way of taking an objec...
  • Quick Start

    How to use Deploy DataStream tasks Deploy the FlinkSql task Task start process How to use The installation of the one-stop platform streampark-console has been introduced in...
  • Development Tools

    5217 2024-06-22 《Apache Accumulo 2.x》
    MiniAccumuloCluster Iterator Test Harness Framework Use Normal Test Outline Limitations Accumulo has several tools that can help developers test their code. MiniAccumuloClu...
  • FATE

    5213 2024-06-22 《Apache Accumulo 2.x》
    Overview REPO Stack FATE Structure in ZooKeeper Administration List/Print Summary (new in 2.1) Cancel Fail Delete Dump Accumulo must implement a number of distributed, m...
  • Vertica

    Description Support Those Engines Using Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key Features Supported DataSource Info Data Type Mapping Source Options ...
  • Snowflake

    Support Those Engines Key Features Description Supported DataSource list Database dependency Data Type Mapping Options tips Task Example simple: CDC(Change data capture) e...
  • Bootstrapping

    5195 2024-06-30 《Apache Hudi 0.15.0》
    Approaches Use Hudi for new partitions alone Convert existing table to Hudi Using Hudi Streamer Using Spark Datasource Writer Using Spark SQL CALL Procedure Using Hudi CLI C...