Devlive 开源社区 本次搜索耗时 0.553 秒,为您找到 1079 个相关结果.
  • Ubuntu 16

    Steps Next Step More Information On a server host that has Internet access, use a command line editor to perform the following: Steps Log in to your host as root . Download...
  • Amazon Linux 2

    步骤 下一步 更多信息 在可以访问 Internet 的服务器主机上,使用命令行编辑器执行以下操作 步骤 以 root 身份登录到您的主机。 将 Ambari 存储库文件下载到安装主机上的目录。 wget - nv https :// username : password@archive . cloudera . com / p /...
  • Iceberg

    4247 2024-07-05 《Apache Kyuubi 1.9.1》
    Iceberg Integration Dependencies Configurations Iceberg Operations Apache Iceberg is an open table format for huge analytic datasets. Iceberg adds tables to compute engines in...
  • 不向前兼容的更新

    不向前兼容的更新 dev 3.2.0 3.0.0 3.2.0 不向前兼容的更新 本文档记录了各版本之间不兼容的更新内容。在升级到相关版本前,请检查本文档。 dev 将mysql驱动版本从8.0.16升级至8.0.33 (#14684 ) 更改了环境变量名称,将 PYTHON_HOME 改为 PYTHON_LAUNCHER , 将 DAT...
  • 告警组件向导

    如何创建告警插件以及告警组 如何创建告警插件以及告警组 在 2.0.0 版本中,用户需要创建告警实例,在创建告警实例时,需要选择告警策略,有三个选项,成功发、失败发,以及成功和失败都发。在执行完工作流或任务时,如果触发告警,调用告警实例发送方法会进行逻辑判断,将告警实例与任务状态进行匹配,匹配则执行该告警实例发送逻辑,不匹配则过滤。创建完告警实例后,需...
  • Google BigQuery

    Iceberg tables Using Iceberg JSON metadata file to create the Iceberg BigLake tables : Steps to add additional configurations to the Hudi writers: Using BigLake Metastore to crea...
  • 远程日志存储

    远程日志存储(Remote Logging) 开启远程日志存储 将任务日志写入 阿里云对象存储(OSS) 将任务日志写入 Amazon S3 将任务日志写入 Google Cloud Storage (GCS) 将任务日志写入 Azure Blob Storage (ABS) 注意事项 远程日志存储(Remote Logging) ...
  • TiDB

    4222 2024-07-05 《Apache Kyuubi 1.9.1》
    TiDB Integration Dependencies Configurations TiDB Operations TiDB is an open-source NewSQL database that supports Hybrid Transactional and Analytical Processing (HTAP) workloa...
  • Setup Options

    More Information The following options are frequently used for Ambari Server setup. -j (or —java-home) Specifies the JAVA_HOME path to use on the Ambari Server and all hosts i...
  • Reliability

    4219 2024-06-29 《Apache Iceberg 1.5.2》
    Concurrent write operations Cost of retries Retry validation Compatibility Iceberg was designed to solve correctness problems that affect Hive tables running in S3. Hive tabl...