Flink Apache Iceberg supports both Apache Flink ‘s DataStream API and Table API. See the Multi-Engine Support page for the integration of Apache Flink. Feature support Flink...
Hive Iceberg supports reading and writing Iceberg tables through Hive by using a StorageHandler . Feature support The following features matrix illustrates the support for diff...
Flink Writes Iceberg support batch and streaming writes With Apache Flink ‘s DataStream API and Table API. Writing with SQL Iceberg support both INSERT INTO and INSERT OVERWRIT...
Creating your first interoperable table Using Apache XTable™ (Incubating) to sync your source tables in different target format involves running sync on your current dataset usi...
Trino just like Presto allows you to query table formats like Hudi, Delta and Iceberg tables using connectors. Users do not need additional configurations to work with OneTable syn...
Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...
Syncing to Hive Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced table on Hive Metastore (HMS). Pre-requisites Source table(s) ...
Configuration Table properties Iceberg tables support table properties to configure table behavior, like the default split size for readers. Read properties Property Defaul...
Presto allows you to query table formats like Hudi, Delta and Iceberg using connectors. The same setup will work for OneTable synced tables as well. For more information and requ...