Devlive 开源社区 本次搜索耗时 0.364 秒,为您找到 1306 个相关结果.
  • Tailwind CSS 部署环境配置

    在将 Tailwind CSS 项目部署到生产环境时,需要做好各项配置和优化工作。本节将详细介绍如何配置和优化部署环境。 环境配置 Docker 配置 # Dockerfile # 构建阶段 FROM node : 16 - alpine as builder WORKDIR / app COPY package *. ...
  • Configuring MySQL for Ranger

    Prerequisites Steps RHEL/CentOS/Oracle/Aamazon Linux SLES Prerequisites When using MySQL, the storage engine used for the Ranger admin policy store tables MUST support trans...
  • Hive CLI

    Hive CLI 综述 Hive CLI任务 VS 连接Hive数据源的SQL任务 创建任务 任务参数 任务样例 Hive CLI任务样例 Hive CLI 综述 使用Hive Cli任务插件 创建Hive Cli 类型的任务执行SQL脚本语句或者SQL任务文件。 执行任务的worker会通过hive -e 命令执行hive SQL脚本语...
  • Object

    4160 2024-06-05 《Ramda 0.29.2》
    clone values eqProps keys omit pick pickAll project prop props keysIn path valuesIn toPairs toPairsIn propOr has hasIn assoc assocPath lens pickBy evolve inv...
  • Apache XTable (incubating)

    4158 2024-06-30 《Apache Hudi 0.15.0》
    Interoperating with XTable Installation Syncing to XTable Hudi Streamer Extensions Hudi (tables created from 0.14.0 onwards) supports syncing to Iceberg and/or Delta Lake with...
  • Apache Paimon (Incubating)

    4157 2024-07-05 《Apache Kyuubi 1.9.1》
    Apache Paimon (Incubating) Integration Dependencies Apache Paimon (Incubating) Operations Apache Paimon (Incubating) is a streaming data lake platform that supports high-speed ...
  • Ubuntu 18

    Steps Next Step More Information On a server host that has Internet access, use a command line editor to perform the following: Steps Log in to your host as root . Download...
  • Apache Spark

    To read a OneTable synced target table (regardless of the table format) in Apache Spark locally or on services like Amazon EMR, Google Cloud’s Dataproc, Azure HDInsight, or Databr...
  • Caching

    4154 2024-06-22 《Apache Accumulo 2.x》
    Configuration Accumulo tablet servers have block caches that buffer data in memory to limit reads from disk. This caching has the following benefits: reduces latency when rea...
  • Procedures

    4142 2025-03-11 《Apache Iceberg 1.8.1》
    Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...