Devlive 开源社区 本次搜索耗时 3.028 秒,为您找到 491 个相关结果.
  • Architecture

    Gobblin Architecture Overview Gobblin Job Flow Gobblin Constructs Source and Extractor Converter Quality Checker Fork Operator Data Writer Data Publisher Gobblin Task Flow...
  • Query

    471 2024-05-25 《Apache JDO 3.2.1》
    Query API JDOQL Methods Result Quick Ref PDF JDOQL Typed API SQL
  • Flink Writes

    Flink Writes Iceberg support batch and streaming writes With Apache Flink ‘s DataStream API and Table API. Writing with SQL Iceberg support both INSERT INTO and INSERT OVERWRIT...
  • SeaTunnel Engine

    about deployment local-mode cluster-mode savepoint checkpoint-storage rest-api tcp engine-jar-storage-mode
  • General

    449 2024-05-25 《Apache JDO 3.2.1》
    Why JDO? Getting Started Specifications TCK API Javadoc JDO v JPA JDO v JPA : API JDO v JPA : ORM License Implementations JDO 3.0 Overview References Glossary
  • Java API

    Iceberg Java API Tables The main purpose of the Iceberg API is to manage table metadata, like schema, partition spec, metadata, and data files that store table data. Table metad...
  • Configuration Glossary

    Table of Contents Properties File Format Creating a Basic Properties File Job Launcher Properties Common Job Launcher Properties SchedulerDaemon Properties CliMRJobLauncher Pr...
  • Configuration

    Spark Configuration Catalogs Spark adds an API to plug in table catalogs that are used to load, create, and manage Iceberg tables. Spark catalogs are configured by setting Spark ...
  • 开发者指南

    Database API Document
  • Writes

    Spark Writes To use Iceberg in Spark, first configure Spark catalogs . Some plans are only available when using Iceberg SQL extensions in Spark 3. Iceberg uses Apache Spark’s D...