Devlive 开源社区 本次搜索耗时 0.613 秒,为您找到 76 个相关结果.
  • 安装后切换到 Oracle 数据库

    关于此任务 条件 步骤 关于此任务 如果在执行初始 HDF 安装或升级后想要将 Oracle 数据库与 SAM 或架构注册表一起使用,则可以切换到 Oracle 数据库。支持 Oracle 数据库 12c 和 11g 第 2 版 条件 您已安装并配置了 Oracle 数据库。 步骤 登录 Ambari Server 并关闭 SAM ...
  • Switching to an Oracle Database After Installation

    About This Task Prerequisites Steps About This Task If you want to use an Oracle database with SAM or Schema Registry after you have performed your initial HDF installation or...
  • Kafka source compatible kafka-connect-json

    How To Use Kafka Sink Mysql Seatunnel connector kafka supports parsing data extracted through kafka connect source, especially data extracted from kafka connect jdbc and kafka c...
  • Avro format

    How To Use Kafka uses example Avro is very popular in streaming data pipeline. Now seatunnel supports Avro format in kafka connector. How To Use Kafka uses example This is a...
  • Configuration

    625 2025-03-14 《Redisson 3.45.0》
    Using Redisson API Programmatic configuration is performed by the Config object instance. For example: Config config = new Config (); config . setTransportMode ( Transpo...
  • AWS

    Iceberg AWS Integrations Iceberg provides integration with different AWS services through the iceberg-aws module. This section describes how to use Iceberg with AWS. Enabling ...
  • Table Services

    380 2024-06-27 《Apache Hudi 0.15.0》
    Bootstrapping Compaction Clustering Metadata Indexing Cleaning Rollback Mechanism Marker Mechanism File Sizing Disaster Recovery
  • Kafka Connect

    Kafka Connect Kafka Connect is a popular framework for moving data in and out of Kafka via connectors. There are many different connectors available, such as the S3 sink for writ...
  • Kafka-HDFS Ingestion

    Getting Started Standalone MapReduce Job Constructs Source and Extractor Writer and Publisher Job Config Properties Metrics and Events Task Level Metrics Task Level Events ...
  • Kafka

    Introduction Pre-requisites Steps Configuration Details What Next? Introduction The Kafka writer allows users to create pipelines that ingest data from Gobblin sources into ...