Devlive 开源社区 本次搜索耗时 0.180 秒,为您找到 1514 个相关结果.
  • Redis

    Description Key features Options host [string] port [int] hash_key_parse_mode [string] keys [string] data_type [string] user [string] auth [string] db_num [int] mode [str...
  • HIVE

    HIVE数据源 使用HiveServer2 使用 HiveServer2 HA Zookeeper 是否原生支持 HIVE数据源 使用HiveServer2 数据源:选择 HIVE 数据源名称:输入数据源的名称 描述:输入数据源的描述 IP 主机名:输入连接 HIVE 的 IP 端口:输入连接 HIVE 的端口 用户名:设置连接 HIVE...
  • Building

    2253 2024-07-05 《Apache Kyuubi 1.9.1》
    Build with Apache Maven Build shaded jar with Apache Maven Build against Different Apache Spark Versions Build against Different Apache Ranger Versions Test with ScalaTest Mave...
  • Configuring MySQL for Ranger

    Prerequisites Steps RHEL/CentOS/Oracle/Aamazon Linux SLES Prerequisites When using MySQL, the storage engine used for the Ranger admin policy store tables MUST support trans...
  • Ubuntu 18

    Steps Next Step More Information On a server host that has Internet access, use a command line editor to perform the following: Steps Log in to your host as root . Download ...
  • Transactions

    2236 2024-05-25 《Apache JDO 3.2.1》
    Locking Non-Transactional When managing the persistence of objects using a PersistenceManager it is common to handle all datastore operations in a transaction. For this reason ...
  • Hive

    Description Key features Options table_name [string] metastore_uri [string] hdfs_site_path [string] read_partitions [list] krb5_path [string] kerberos_principal [string] ke...
  • Marker Mechanism

    2234 2024-06-30 《Apache Hudi 0.15.0》
    Purpose of Markers Marker structure Marker Writing Options Direct Write Markers Timeline Server Markers (Default) Marker Configuration Parameters Purpose of Markers A write...
  • 为 HDF 服务安装数据库

    安装和配置MySQL 安装和配置 Postgres 使用 Oracle 数据库 安装 Schema Registry、SAM、Druid 和 Superset 时,您需要一个关系数据存储来存储元数据。您可以使用 MySQL、Postgres、Oracle 或 MariaDB。这些主题描述如何安装 MySQL、Postgres 和 Oracle 以及如...
  • AWS S3

    2231 2024-07-01 《Apache Hudi 0.15.0》
    AWS configs AWS Credentials AWS Libs AWS S3 Versioned Bucket In this page, we explain how to get your Hudi spark job to store into AWS S3. AWS configs There are two configur...