Devlive 开源社区 本次搜索耗时 0.242 秒,为您找到 1369 个相关结果.
  • Socket

    Support Those Engines Key features Description Data Type Mapping Options How to Create a Socket Data Synchronization Jobs Socket source connector Support Those Engines ...
  • EventMesh Stream

    EventMesh Stream Overview of Event Streaming Requirements Functional Requirements Design Details Architecture Design EventMesh-Stream Component Event Event Channel Event End...
  • Confirm Hosts

    Next Step Confirm Hosts prompts you to confirm that Ambari has located the correct hosts for your cluster and to check those hosts to make sure they have the correct directories...
  • EventMesh Workflow

    EventMesh Workflow Business Problem CNCF Serverless Workflow EventMesh Workflow AsyncAPI Workflow Example EventMesh Workflow Engine EventMesh Catalog Design EventMesh Workflo...
  • Confirm Hosts

    Next Step Confirm Hosts prompts you to confirm that Ambari has located the correct hosts for your cluster and to check those hosts to make sure they have the correct directories...
  • Kyuubi v.s. HiveServer2

    3422 2024-07-05 《Apache Kyuubi 1.9.1》
    Introduction Hive on Spark Differences Between Kyuubi and HiveServer2 Performance References Introduction HiveServer2 is a service that enables clients to execute Hive QL qu...
  • Microservices integration

    3420 2025-03-15 《Redisson 3.45.0》
    Spring Boot For Spring Boot usage please refer to Spring Boot article. Micronaut Redisson integrates with Micronaut framework. Supports Micronaut 2.0.x - 4.x.x Usage 1. A...
  • BigLake Metastore

    Syncing to BigLake Metastore This document walks through the steps to register an Apache XTable™ (Incubating) synced Iceberg table in BigLake Metastore on GCP. Pre-requisites S...
  • CloudEvents Integration

    CloudEvents Integration Introduction Requirements Functional Requirements Performance Requirements Design Details Pluggable Protocols Process of CloudEvents under EventMesh Fo...
  • DDL

    3396 2025-03-11 《Apache Iceberg 1.8.1》
    Spark DDL To use Iceberg in Spark, first configure Spark catalogs . Iceberg uses Apache Spark’s DataSourceV2 API for data source and catalog implementations. CREATE TABLE Spark...