Devlive 开源社区 本次搜索耗时 0.241 秒,为您找到 839 个相关结果.
  • 4. Kyuubi Release Guide

    Kyuubi Release Guide Introduction Overview Decide to release Checklist to proceed to the next step Prepare for the release One-time setup instructions ASF authentication Subve...
  • 9.6 Custom blocks (*)

    1292 2024-05-23 《R Markdown Cookbook》
    Syntax Adding a shaded box Including icons Section 2.7 of the bookdown book mentioned how we can use custom blocks in R Markdown to customize the appearance of blocks of cont...
  • Architecture

    1292 2024-07-05 《Apache Kyuubi 1.9.1》
    Introduction Architecture Overview Unified Interface Runtime Resource Resiliency High Availability & Load Balance Authentication & Authorization Conclusions Introduction K...
  • Sqoop

    SQOOP 节点 综述 创建任务 任务参数 任务样例 配置 Sqoop 环境 配置 Sqoop 任务节点 查看运行结果 SQOOP 节点 综述 SQOOP 任务类型,用于执行 SQOOP 程序。对于 SQOOP 节点,worker 会通过执行 sqoop 命令来执行 SQOOP 任务。 创建任务 点击项目管理 -> 项目名称 -...
  • Docker Builds

    1284 2024-05-24 《Apache Superset 4.0.1》
    Key Image Tags and Examples Caching About database drivers On supporting arm64 AND amd64 Working with Apple silicon The Apache Superset community extensively uses Docker for ...
  • JDBC

    Description Using Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key Features Options driver [string] user [string] password [string] url [string] query [stri...
  • Hive

    1282 2024-06-29 《Apache Iceberg 1.5.2》
    Feature support Enabling Iceberg support in Hive Hive 4.0.0-beta-1 Hive 4.0.0-alpha-2 Hive 4.0.0-alpha-1 Hive 2.3.x, Hive 3.1.x Loading runtime jar Enabling support Hadoop con...
  • Error Quick Reference Manual

    SeaTunnel API Error Codes SeaTunnel Common Error Codes Assert Connector Error Codes Cassandra Connector Error Codes Slack Connector Error Codes MyHours Connector Error Codes ...
  • S3File

    Support Those Engines Key Features Description Supported DataSource Info Dependency Data Type Mapping JSON File Type Text Or CSV File Type Orc File Type Parquet File Type ...
  • Creating your first interoperable table

    Pre-requisites Steps Initialize a pyspark shell Create dataset Running sync Conclusion Next steps Using OneTable to sync your source tables in different target format invo...