Devlive 开源社区 本次搜索耗时 0.787 秒,为您找到 172 个相关结果.
  • REST API v1

    9373 2024-07-05 《Apache Kyuubi 1.9.1》
    REST API v1 Session Resource GET /sessions Response Body GET /sessions/${sessionHandle} Response Body GET /sessions/${sessionHandle}/info/${infoType} Request Parameters Respon...
  • Flink Getting Started

    9336 2024-06-29 《Apache Iceberg 1.5.2》
    Preparation when using Flink SQL Client Flink’s Python API Adding catalogs. Catalog Configuration Hive catalog Creating a table Writing Branch Writes Reading Type conversi...
  • 2. Getting Started With Kyuubi and DBeaver

    Getting Started With Kyuubi and DBeaver What is DBeaver Preparation Get DBeaver and Install Get Kyuubi Started Configurations Start DBeaver Select a database Edit the Driver ...
  • AWS

    8933 2025-03-11 《Apache Iceberg 1.8.1》
    Iceberg AWS Integrations Iceberg provides integration with different AWS services through the iceberg-aws module. This section describes how to use Iceberg with AWS. Enabling ...
  • Deployment

    8915 2024-06-26 《Apache Amoro 0.6.1》
    System requirements Download the distribution Source code compilation Configuration Configure the service address Configure system database Configure high availability Config...
  • MongoDB

    Support Those Engines Key Features Description Supported DataSource Info Data Type Mapping Source Options Tips How to Create a MongoDB Data Synchronization Jobs Parameter I...
  • Getting Started

    8731 2024-07-05 《Apache Kyuubi 1.9.1》
    Requirements Installation Install Kyuubi Install Spark Configuration Start Kyuubi Operate Clients Open Connections Execute Statements Start Engines Close Connections Stop...
  • Deployment

    8614 2024-07-01 《Apache Hudi 0.15.0》
    Deploying Hudi Streamer Spark Datasource Writer Jobs Upgrading Downgrading Migrating This section provides all the help you need to deploy and operate Hudi tables at scale. ...
  • Postgre CDC

    Support Those Engines Key features Description Supported DataSource Info Using Dependency Install Jdbc Driver For Spark/Flink Engine For SeaTunnel Zeta Engine Data Type Map...
  • Docker 部署

    前置条件 1. 安装 docker 2. 安装 docker-compose 部署 Apache StreamPark™ 1. 基于 h2 和 docker-compose 部署 Apache StreamPark™ 2. 部署 3. 配置flink home 4. 配置session集群 5. 提交 Flink 作业 使用已有的 Mysql ...