Devlive 开源社区 本次搜索耗时 1.181 秒,为您找到 1139 个相关结果.
  • Docker Demo

    3737 2024-06-28 《Apache Hudi 0.15.0》
    A Demo using Docker containers Prerequisites Setting up Docker Cluster Build Hudi Bringing up Demo Cluster Demo Step 1 : Publish the first batch to Kafka Step 2: Incrementall...
  • 3. Integration with Hive Metastore

    Integration with Hive Metastore Requirements Default Behavior Related Configurations Remote Metastore Database Remote Metastore Server Activate Configurations Via kyuubi-defau...
  • JDBC Connector

    JDBC 信息配置 semantic 语义配置 EXACTLY_ONCE AT_LEAST_ONCE && NONE 其他配置 JDBC 读取数据 queryFunc获取一条sql resultFunc 处理查询到的数据 JDBC 读取写入 根据数据流生成目标SQL 设置写入批次大小 多实例 JDBC 支持 手动指定 JDBC 连接信...
  • Flink K8s 集成支持

    额外环境要求 集成准备 Kubernetes 连接配置 Kubernetes RBAC 配置 Docker 远程容器服务配置 任务提交 Application 任务发布 Session 任务发布 相关参数配置 StreamPark Flink Kubernetes 基于 Flink Native Kubernetes 实现,支持以下 F...
  • Spark Quick Start

    3592 2024-06-28 《Apache Hudi 0.15.0》
    Setup Spark 3 Support Matrix Spark Shell/SQL Setup project Create Table Insert data Query data Update data Merging Data Delete data Time Travel Query Incremental query ...
  • Quick Start

    Quick Start Configure FLINK_HOME Configure Flink Cluster Start Flink Standalone Session on Server Configure Flink Cluster Create Job Main Parameters Create Job Save Job Bui...
  • Contributing to Superset

    3585 2024-05-25 《Apache Superset 4.0.1》
    Contributing to Superset Orientation Types of Contributions Report Bug Submit Ideas or Feature Requests Fix Bugs Implement Features Improve Documentation Add Translations As...
  • Writing Tables

    3525 2024-07-01 《Apache Hudi 0.15.0》
    What are some ways to write a Hudi table? How is a Hudi writer job deployed? Can I implement my own logic for how input records are merged with record on storage? How do I delet...
  • Replication

    3516 2024-06-22 《Apache Accumulo 2.x》
    Overview Configuration Site Configuration Instance Configuration Table Configuration Monitoring Work Assignment ReplicaSystems AccumuloReplicaSystem Other Configuration E...
  • Hadoop 资源集成

    在 Flink on Kubernetes 上使用 Apache Hadoop 资源 1. Apache HDFS 1.1 添加 shade jar 1.2、添加 core-site.xml 和 hdfs-site.xml 2、Apache Hive i、添加 hive 相关的 jar 2.1. 添加 hive 的配置文件(hive-site.x...