Devlive 开源社区 本次搜索耗时 0.615 秒,为您找到 872 个相关结果.
  • Parquet HDFS

    Description Usage Example Pipeline Configuration Configuration Developer Notes Description An extension to FsDataWriter that writes in Parquet format in the form of either...
  • Basics

    1. Deploy Kyuubi engines on Yarn 2. Deploy Kyuubi engines on Kubernetes 3. Integration with Hive Metastore 4. Kyuubi High Availability Guide
  • Flink

    825 2024-06-26 《Apache Amoro 0.6.1》
    Flink Getting Started Flink DDL Flink DML Flink DataStream Using Logstore
  • Performance

    Performance Iceberg is designed for huge tables and is used in production where a single table can contain tens of petabytes of data. Even multi-petabyte tables can be read from ...
  • Run Java SDK Demo

    Run Java SDK Demo 1. TCP 1.1 ASYNC 1.2 BROADCAST 2 HTTP 2.1 ASYNC 3 GRPC 3.1 ASYNC Publish & Webhook Subscribe 3.2 SYNC Publish & Stream Subscribe 3.3 Publish BATCH Message ...
  • Presto

    Querying from Presto Presto allows you to query table formats like Hudi, Delta and Iceberg using connectors. The same setup will work for Apache XTable™ (Incubating) synced tables...
  • Trino

    Querying from Trino Trino just like Presto allows you to query table formats like Hudi, Delta and Iceberg tables using connectors. Users do not need additional configurations to w...
  • High Level Consumer

    Problem Statement New Design & Details Problem Statement Current Gobblin Kafka High Level Consumer uses Kafka Consumer (0.8) APIs and gobblin support for them will be depreca...
  • Contributing

    Contributing to Gobblin Code Contributions Documentation Contributions Contributing to Gobblin You can contribute to Gobblin in multiple ways. For resources and guides, please...
  • Knative

    Knative Prerequisite Create Knative Source and Sink Set up EventMesh Configuration Step 2: Publish an Event Message from Knative Step 3: Subscribe from an EventMesh Expected Res...