Devlive 开源社区 本次搜索耗时 0.470 秒,为您找到 804 个相关结果.
  • Java Quickstart

    3729 2024-06-29 《Apache Iceberg 1.5.2》
    Create a table Using a Hive catalog Using a Hadoop catalog Branching and Tagging Creating branches and tags Committing to branches Reading from branches and tags Replacing an...
  • Ogg Format

    Format Options How to use Ogg format Kafka uses example Oracle GoldenGate (a.k.a ogg) is a managed service providing a real-time data mesh platform, which uses replication to ...
  • 3. Trouble Shooting

    Trouble Shooting Common Issues java.lang.UnsupportedClassVersionError .. Unsupported major.minor version 52.0 org.apache.spark.SparkException: When running with master ‘yarn’ eith...
  • Error Quick Reference Manual

    SeaTunnel API Error Codes SeaTunnel Common Error Codes Assert Connector Error Codes Cassandra Connector Error Codes Slack Connector Error Codes MyHours Connector Error Codes ...
  • OneSignal

    Description Key features Options url [String] password [String] method [String] params [Map] body [String] poll_interval_millis [int] retry [int] retry_backoff_multiplier_...
  • REST API v1

    3674 2024-07-05 《Apache Kyuubi 1.9.1》
    REST API v1 Session Resource GET /sessions Response Body GET /sessions/${sessionHandle} Response Body GET /sessions/${sessionHandle}/info/${infoType} Request Parameters Respon...
  • OssFile

    Support Those Engines Usage Dependency For Spark/Flink Engine For SeaTunnel Zeta Engine Key features Data Type Mapping Orc File Type Parquet File Type Options path [string]...
  • Hive

    3650 2024-06-29 《Apache Iceberg 1.5.2》
    Feature support Enabling Iceberg support in Hive Hive 4.0.0-beta-1 Hive 4.0.0-alpha-2 Hive 4.0.0-alpha-1 Hive 2.3.x, Hive 3.1.x Loading runtime jar Enabling support Hadoop con...
  • 2.2 R Markdown anatomy

    3643 2024-05-09 《R Markdown Cookbook》
    YAML metadata Narrative Code chunks Document body References We can dig one level deeper by considering the different components of an R Markdown. Specifically, let’s look at...
  • Creating your first interoperable table

    Pre-requisites Steps Initialize a pyspark shell Create dataset Running sync Conclusion Next steps Using OneTable to sync your source tables in different target format invo...