Devlive 开源社区 本次搜索耗时 0.560 秒,为您找到 282 个相关结果.
  • Socket

    Support Those Engines Key features Description Data Type Mapping Options How to Create a Socket Data Synchronization Jobs Socket source connector Support Those Engines ...
  • K8S PVC 资源使用

    K8s PVC 资源使用说明 K8s PVC 资源使用说明 当前版本 StreamPark Flink-K8s 任务对 PVC 资源(挂载 checkpoint/savepoint/logs 等文件资源)的支持基于 pod-template。 Native-Kubernetes Session 由创建 Session Cluster 时控制,这里不再...
  • Dell

    3326 2025-03-11 《Apache Iceberg 1.8.1》
    Iceberg Dell Integration Dell ECS Integration Iceberg can be used with Dell’s Enterprise Object Storage (ECS) by using the ECS catalog since 0.15.0. See Dell ECS for more infor...
  • Socket

    Support Those Engines Key features Description Sink Options Task Example Changelog 2.2.0-beta 2022-09-26 Socket sink connector Support Those Engines Spark Flink SeaTu...
  • Debugging Kyuubi

    3302 2024-07-05 《Apache Kyuubi 1.9.1》
    Debugging Server Debugging Engine Flink Engine Trino Engine Hive Engine Debugging Apps Spark Engine Flink Engine You can use the Java Debug Wire Protocol to debug Kyuubi ...
  • Flink Actions

    3266 2024-06-29 《Apache Iceberg 1.5.2》
    Rewrite files action. Rewrite files action. Iceberg provides API to rewrite small files into large files by submitting Flink batch jobs. The behavior of this Flink action is the...
  • 任务类型

    Appendix Shell SubProcess Dynamic Dependent Stored Procedure SQL Spark MapReduce Python Flink HTTP DataX Sqoop Pigeon Conditions Switch SeaTunnel Amazon EMR Apac...
  • Console

    Support Connector Version Support Those Engines Description Key Features Options Task Example Simple: Multiple Sources Simple: Console Sample Data Console sink connector...
  • Java Custom Catalog

    3088 2025-03-14 《Apache Iceberg 1.8.1》
    Custom Catalog It’s possible to read an iceberg table either from an hdfs path or from a hive table. It’s also possible to use a custom metastore in place of hive. The steps to do...
  • 变量管理

    背景介绍 创建变量 在Flink SQL中引用变量 在Flink JAR作业的args中引用变量 背景介绍 在实际生产环境中,Flink作业一般很复杂,会依赖多个外部组件,例如,从Kafka中消费数据时要从HBase或Redis中去获取相关数据,然后将关联好的数据写入到外部组件,这样的情况下会导致如下问题: Flink作业想要关联这些组件,需...