Devlive 开源社区 本次搜索耗时 0.653 秒,为您找到 160 个相关结果.
  • HTTP Connector

    HTTP 异步写入 Apache StreamPark™ 方式写入 http异步写入支持类型 http异步写入配置参数列表 http异步写入数据 其他配置 一些后台服务通过 HTTP 请求接收数据,这种场景下 Apache Flink 可以通过 HTTP 请求写入结果数据,目前 Apache Flink 官方未提供通过 HTTP 请求写入 数据...
  • Flink Quick Start

    1598 2024-06-28 《Apache Hudi 0.15.0》
    Setup Flink Support Matrix Download Flink and Start Flink cluster Start Flink SQL client Create Table Insert Data Query Data Update Data Delete Data Row-level Delete Batch...
  • Clustering

    1577 2024-06-30 《Apache Hudi 0.15.0》
    Background How is compaction different from clustering? Clustering Architecture Overall, there are 2 steps to clustering Schedule clustering Execute clustering Clustering Use...
  • Design & Concepts

    1566 2024-07-01 《Apache Hudi 0.15.0》
    How does Hudi ensure atomicity? Does Hudi extend the Hive table layout? What concurrency control approaches does Hudi adopt? Hudi’s commits are based on transaction start time i...
  • 2.2. How To Use Spark Adaptive Query Execution (AQE) in Kyuubi

    How To Use Spark Adaptive Query Execution (AQE) in Kyuubi The Basics of AQE Dynamically Switch Join Strategies Dynamically Coalesce Shuffle Partitions Other Tips for Best Practise...
  • Troubleshooting

    1527 2024-07-01 《Apache Hudi 0.15.0》
    Writing Tables org.apache.parquet.io.InvalidRecordException: Parquet/Avro schema mismatch: Avro field ‘col1’ not found java.lang.UnsupportedOperationException: org.apache.parquet....
  • HTTP Connector

    http asynchronous write Write with Apache StreamPark™ http asynchronous write support type Configuration list of HTTP asynchronous write HTTP writes data asynchronously Other ...
  • Creating Your First Dashboard

    1512 2024-05-25 《Apache Superset 4.0.1》
    Connecting to a new database ​ Registering a new table ​ Customizing column properties ​ Superset semantic layer ​ Creating charts in Explore view ​ Creating a slice and ...
  • Deployment

    1465 2024-07-01 《Apache Hudi 0.15.0》
    Deploying Hudi Streamer Spark Datasource Writer Jobs Upgrading Downgrading Migrating This section provides all the help you need to deploy and operate Hudi tables at scale. ...
  • Hadoop Resource Integration

    Using Apache Hadoop resource in Flink on Kubernetes 1. Apache HDFS 1.1 Add the shaded jar 1.2. add core-site.xml and hdfs-site.xml 2. Apache Hive 2.1. Add Hive-related jars 2...