Actions States Active and Archived timeline LSM Timeline Archival Configs Spark write client configs Flink Options At its core, Hudi maintains a timeline which is a log of...
Hive Iceberg supports reading and writing Iceberg tables through Hive by using a StorageHandler . Feature support The following features matrix illustrates the support for diff...
Requirements Usage DB-API Use PyHive with Pandas Authentication PyHive is a collection of Python DB-API and SQLAlchemy interfaces for Hive. PyHive can connect with the Kyuub...
Catalog Configuration Runtime configuration Read options Write options Catalog Configuration A catalog is created and named by executing the following query (replace <catalog...
Introduction Supported table format Supported column data type How to use Optimize history data Syntax Examples Optimize incremental data To improve query speed, Kyuubi su...
Support Connector Version Support Those Engines Key Features Description Supported DataSource Info Database Dependency Data Type Mapping Source Options Tips Task Example S...
Apache Paimon (Incubating) Integration Dependencies Configurations Apache Paimon (Incubating) Operations Apache Paimon(incubating) is a streaming data lake platform that suppo...
Instructions Install Hive JDBC JDBC URL Example Instructions Kyuubi is fully compatible with Hive JDBC and ODBC drivers that let you connect to popular Business Intelligence ...
Support Those Engines Description Using Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key Features Supported DataSource Info Data Type Mapping Sink Options Ti...
Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...