Kafka Connect Kafka Connect is a popular framework for moving data in and out of Kafka via connectors. There are many different connectors available, such as the S3 sink for writ...
Creating your first interoperable table Using Apache XTable™ (Incubating) to sync your source tables in different target format involves running sync on your current dataset usi...
Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...
Flink Writes Iceberg support batch and streaming writes With Apache Flink ‘s DataStream API and Table API. Writing with SQL Iceberg support both INSERT INTO and INSERT OVERWRIT...
Syncing to Hive Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced table on Hive Metastore (HMS). Pre-requisites Source table(s) ...
Iceberg Java API Tables The main purpose of the Iceberg API is to manage table metadata, like schema, partition spec, metadata, and data files that store table data. Table metad...
Spark Configuration Catalogs Spark adds an API to plug in table catalogs that are used to load, create, and manage Iceberg tables. Spark catalogs are configured by setting Spark ...