Trino just like Presto allows you to query table formats like Hudi, Delta and Iceberg tables using connectors. Users do not need additional configurations to work with OneTable syn...
Flink Apache Iceberg supports both Apache Flink ‘s DataStream API and Table API. See the Multi-Engine Support page for the integration of Apache Flink. Feature support Flink...
Kafka Connect Kafka Connect is a popular framework for moving data in and out of Kafka via connectors. There are many different connectors available, such as the S3 sink for writ...
Hive Iceberg supports reading and writing Iceberg tables through Hive by using a StorageHandler . Feature support The following features matrix illustrates the support for diff...
Hudi and Iceberg tables Delta Lake table To read a OneTable synced target table (regardless of the table format) in Amazon Redshift, users have to create an external schema and ...
Presto allows you to query table formats like Hudi, Delta and Iceberg using connectors. The same setup will work for OneTable synced tables as well. For more information and requ...
Creating your first interoperable table Using Apache XTable™ (Incubating) to sync your source tables in different target format involves running sync on your current dataset usi...
Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...
Flink Writes Iceberg support batch and streaming writes With Apache Flink ‘s DataStream API and Table API. Writing with SQL Iceberg support both INSERT INTO and INSERT OVERWRIT...
Syncing to Hive Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced table on Hive Metastore (HMS). Pre-requisites Source table(s) ...