Devlive 开源社区 本次搜索耗时 0.575 秒,为您找到 355 个相关结果.
  • S3File

    Support Those Engines Key Features Description Supported DataSource Info Database Dependency Data Type Mapping Orc File Type Parquet File Type Sink Options path [string] h...
  • 告警通知

    webapi.conf server.conf 夜莺告警通知,内置支持邮件、钉钉机器人、企微机器人、飞书机器人多种方式作为发送通道,也支持调用自定义脚本和Webhook,给用户自定义发送通道的能力。相关配置在 webapi.conf 和 server.conf 中都有涉及。这里分别讲解。 webapi.conf [[ NotifyChannels ...
  • General

    4009 2024-07-01 《Apache Hudi 0.15.0》
    When is Hudi useful for me or my organization? What are some non-goals for Hudi? What is incremental processing? Why does Hudi docs/talks keep talking about it? How is Hudi opti...
  • Creating GitHub CLI extensions

    4002 2024-05-07 《GitHub Cli》
    About GitHub CLI extensions Creating an interpreted extension with gh extension create Creating a precompiled extension in Go with gh extension create Creating a non-Go precom...
  • CLI

    3913 2024-07-01 《Apache Hudi 0.15.0》
    Local set up Hudi CLI Bundle setup Using hudi-cli Inspecting Commits Drilling Down to a specific Commit FileSystem View Statistics Archived Commits Compactions Validate Com...
  • Troubleshooting

    3910 2024-07-01 《Apache Hudi 0.15.0》
    Writing Tables org.apache.parquet.io.InvalidRecordException: Parquet/Avro schema mismatch: Avro field ‘col1’ not found java.lang.UnsupportedOperationException: org.apache.parquet....
  • Utility

    3812 2024-06-07 《Underscore.js 1.13.6》
    noConflict identity constant noop times random mixin iteratee uniqueId escape unescape result now template noConflict _.noConflict() source Give control of the...
  • MongoDB

    Support Those Engines Key Features Description Supported DataSource Info Data Type Mapping Source Options Tips How to Create a MongoDB Data Synchronization Jobs Parameter I...
  • Translations & Localization

    3799 2024-06-03 《Docusaurus 1.14.7》
    Docusaurus Translation Configurations Translating Your Existing Docs Enabling Translations on Pages Gathering Strings to Translate Custom Translation Strings How Strings Get T...
  • Hadoop Resource Integration

    Using Apache Hadoop resource in Flink on Kubernetes 1. Apache HDFS 1.1 Add the shaded jar 1.2. add core-site.xml and hdfs-site.xml 2. Apache Hive 2.1. Add Hive-related jars 2...