Devlive 开源社区 本次搜索耗时 0.603 秒,为您找到 315 个相关结果.
  • Kafka

    Introduction Pre-requisites Steps Configuration Details What Next? Introduction The Kafka writer allows users to create pipelines that ingest data from Gobblin sources into ...
  • Monitoring

    Overview Metrics Collecting and Reporting Metrics Reporting Metrics collection JVM Metrics Pre-defined Job Execution Metrics Job Execution History Store Email Notifications ...
  • 开发环境

    Flink K8s 集成支持 K8S PVC 资源使用 Hadoop 资源集成
  • Java Custom Catalog

    Custom Catalog It’s possible to read an iceberg table either from an hdfs path or from a hive table. It’s also possible to use a custom metastore in place of hive. The steps to do...
  • Flink Connector

    Flink Connector Apache Flink supports creating Iceberg table directly without creating the explicit Flink catalog in Flink SQL. That means we can just create an iceberg table by s...
  • Queries

    Spark Queries To use Iceberg in Spark, first configure Spark catalogs . Iceberg uses Apache Spark’s DataSourceV2 API for data source and catalog implementations. Querying with S...
  • Deployment

    Gobblin Execution Modes Overview One important feature of Gobblin is that it can be run on different platforms. Currently, Gobblin can run in standalone mode (which runs on a sing...
  • Flink DDL

    DDL commands CREATE Catalog Hive catalog This creates an Iceberg catalog named hive_catalog that can be configured using 'catalog-type'='hive' , which loads tables from Hive m...
  • Deployment

    Flink on K8s Kubernetes PVC Resource usage Hadoop Resource Integration
  • Docker Integration

    Introduction Docker Docker Repositories Run the docker image with simple wikipedia jobs Use Gobblin Standalone on Docker for Kafka and HDFS Ingestion Run Gobblin as a Service ...