Documentation Apache Iceberg is an open table format for huge analytic datasets. Iceberg adds tables to compute engines including Spark, Trino, PrestoDB, Flink, Hive and Impala u...
Hudi Integration Dependencies Hudi Operations Apache Hudi (pronounced “hoodie”) is the next generation streaming data lake platform. Apache Hudi brings core warehouse and datab...
Spark Structured Streaming Iceberg uses Apache Spark’s DataSourceV2 API for data source and catalog implementations. Spark DSv2 is an evolving API with different levels of support...
Java API Quickstart Create a table Tables are created using either a Catalog or an implementation of the Tables interface. Using a Hive catalog The Hive catalog connects to...
Running the TCK Demonstrating Compliance In order to demonstrate compliance with the Java Data Objects specification, an implementation must pass all of the tests in the Technol...
Managing Watermarks in a Job Basics Task Failures Multi-Dataset Jobs Gobblin State Deep Dive State class hierarchy How States are Used in a Gobblin Job This page has two p...
Hudi Integration Dependencies Configurations Hudi Operations Apache Hudi (pronounced “hoodie”) is the next generation streaming data lake platform. Apache Hudi brings core war...
Introduction Pre-requisites Steps Configuration Details What Next? Introduction The Kafka writer allows users to create pipelines that ingest data from Gobblin sources into ...
Installing and Configuring MySQL Installing and Configuring Postgres Using an Oracle Database When installing Schema Registry, SAM, Druid, and Superset, you require a relationa...