Devlive 开源社区 本次搜索耗时 0.559 秒,为您找到 1079 个相关结果.
  • Amazon Athena

    To read a OneTable synced target table (regardless of the table format) in Amazon Athena, you can create the table either by: Using a DDL statement as mentioned in the following...
  • 6. Installing, Configuring, and Deploying a Cluster

    Use the Ambari Cluster Install Wizard running in your browser to install, configure, and deploy your cluster, as follows: Start the Ambari Server Log In to Apache Ambari Launc...
  • JsonPath

    Description Options common options [string] fields[array] option src_field dest_field dest_type path Read Json Example Read SeatunnelRow Example Changelog JsonPath tr...
  • Metrics Reporting

    1898 2025-03-10 《Apache Iceberg 1.8.1》
    Metrics Reporting As of 1.1.0 Iceberg supports the MetricsReporter and the MetricsReport APIs. These two APIs allow expressing different metrics reports while supporting a plu...
  • Hive Beeline

    1897 2024-07-05 《Apache Kyuubi 1.9.1》
    Prerequisites Kyuubi supports Apache Hive beeline that works with Kyuubi server. Hive beeline is a SQLLine CLI based on the Hive JDBC Driver . Prerequisites Kyuubi server ins...
  • JDBC

    1889 2025-03-25 《Apache Gobblin 0.17.0》
    Gobblin JDBC writer & publisher Proposed design Requirements Design summary Design detail WriterInitializer JdbcWriterInitializer AvroFieldsPickConverter AvroToJdbcEntryConve...
  • BigLake Metastore

    Syncing to BigLake Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced Iceberg table in BigLake Metastore on GCP. Pre-requisites S...
  • Setting Up a Local Repository

    Based on your Internet access, choose one of the following options: No Internet Access This option involves downloading the repository tarball, moving the tarball to the sele...
  • DDL

    1866 2025-03-11 《Apache Iceberg 1.8.1》
    Spark DDL To use Iceberg in Spark, first configure Spark catalogs . Iceberg uses Apache Spark’s DataSourceV2 API for data source and catalog implementations. CREATE TABLE Spark...
  • Pigeon

    Pigeon 创建任务 任务参数 Pigeon Pigeon任务类型是通过调用远程websocket服务,实现远程任务的触发,状态、日志的获取,是 DolphinScheduler 通用远程 websocket 服务调用任务 创建任务 拖动工具栏中的任务节点到画板中即能完成任务创建 任务参数 默认参数说明请参考DolphinSchedul...