Usage Manage kyuubi servers List server Create server Get server Delete server Manage kyuubi engines List engine Get engine Delete engine Usage bin / kyuubi - ctl -- h...
Custom Catalog It’s possible to read an iceberg table either from an hdfs path or from a hive table. It’s also possible to use a custom metastore in place of hive. The steps to do...
Evolution Iceberg supports in-place table evolution . You can evolve a table schema just like SQL — even in nested structures — or change partition layout when data volume chang...
Running Tests Fully Running Tests for a Module Running Tests for a Single Test Kyuubi can be tested based on Apache Maven and the ScalaTest Maven Plugin, please refer to the ...
The Engine Configuration Guide The Engine Configuration Guide Kyuubi aims to bring Spark to end-users who need not qualify with Spark or something else related to the big data a...
Daft Daft is a distributed query engine written in Python and Rust, two fast-growing ecosystems in the data engineering and machine learning industry. It exposes its flavor of t...
Aliyun OSS configs Aliyun OSS Credentials Aliyun OSS Libs In this page, we explain how to get your Hudi spark job to store into Aliyun OSS. Aliyun OSS configs There are two c...
Querying from Apache Spark To read an Apache XTable™ (Incubating) synced target table (regardless of the table format) in Apache Spark locally or on services like Amazon EMR, Goog...
Spark Streaming Spark Streaming You can write Hudi tables using spark’s structured streaming. Scala // spark-shell // prepare to stream write to new table import org ....