Spark Configuration Catalogs Spark adds an API to plug in table catalogs that are used to load, create, and manage Iceberg tables. Spark catalogs are configured by setting Spark ...
Scan planning Metadata filtering Data filtering Iceberg is designed for huge tables and is used in production where a single table can contain tens of petabytes of data. Even ...
Event Logging StatsD Logging Event Logging Superset by default logs special action events in its internal database (DBEventLogger). These logs can be accessed on the UI by navi...
Spark Writes To use Iceberg in Spark, first configure Spark catalogs . Some plans are only available when using Iceberg SQL extensions in Spark 3. Iceberg uses Apache Spark’s D...
Flink Queries Iceberg support streaming and batch read With Apache Flink ‘s DataStream API and Table API. Reading with SQL Iceberg support both streaming and batch read in Flink...
Cypher() Cypher in an Expression SELECT Clause Cypher queries are constructed using a function called cypher in ag_catalog which returns a Postgres SETOF records . Cypher() c...