Devlive 开源社区 本次搜索耗时 0.768 秒,为您找到 172 个相关结果.
  • Hive Metastore

    Syncing to Hive Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced table on Hive Metastore (HMS). Pre-requisites Source table(s) ...
  • Configuration

    3225 2025-03-10 《Apache Iceberg 1.8.1》
    Configuration Table properties Iceberg tables support table properties to configure table behavior, like the default split size for readers. Read properties Property Defaul...
  • Hudi

    3181 2024-07-05 《Apache Kyuubi 1.9.1》
    Hudi Integration Dependencies Configurations Hudi Operations Apache Hudi (pronounced “hoodie”) is the next generation streaming data lake platform. Apache Hudi brings core war...
  • Solution for Big Result Sets

    3130 2024-07-05 《Apache Kyuubi 1.9.1》
    Incremental collection Use in single connections Change incremental collection mode in session Typically, when a user submits a SELECT query to Spark SQL engine, the Driver cal...
  • Spark Configuration

    3117 2024-06-26 《Apache Amoro 0.6.1》
    Catalogs configuration Using Mixed-Format in a standalone catalog Using Mixed-Format in session catalog The high availability configuration Catalogs configuration Using Mixe...
  • IBM Cloud

    3114 2024-07-01 《Apache Hudi 0.15.0》
    IBM COS configs IBM Cloud Object Storage Credentials IBM Cloud Object Storage Libs In this page, we explain how to get your Hudi spark job to store into IBM Cloud Object Storag...
  • Client Properties (3.x)

    3021 2024-06-22 《Apache Accumulo 2.x》
    Below are properties set in accumulo-client.properties that configure Accumulo clients . All properties have been part of the API since 2.0.0 (unless otherwise specified): Pr...
  • K8S PVC 资源使用

    K8s PVC 资源使用说明 K8s PVC 资源使用说明 当前版本 StreamPark Flink-K8s 任务对 PVC 资源(挂载 checkpoint/savepoint/logs 等文件资源)的支持基于 pod-template。 Native-Kubernetes Session 由创建 Session Cluster 时控制,这里不再...
  • Configuration

    2861 2025-03-11 《Apache Iceberg 1.8.1》
    Spark Configuration Catalogs Spark adds an API to plug in table catalogs that are used to load, create, and manage Iceberg tables. Spark catalogs are configured by setting Spark ...
  • Writes

    2736 2025-03-11 《Apache Iceberg 1.8.1》
    Spark Writes To use Iceberg in Spark, first configure Spark catalogs . Some plans are only available when using Iceberg SQL extensions in Spark 3. Iceberg uses Apache Spark’s D...