Devlive 开源社区 本次搜索耗时 0.610 秒,为您找到 160 个相关结果.
  • AWS Datasync

    DataSync 节点 综述 创建任务 任务样例 独有参数 环境配置 DataSync 节点 综述 AWS DataSync 是一种在线数据传输服务,可简化、自动化和加速本地存储系统和 AWS Storage 服务之间,以及不同 AWS Storage 服务之间的数据移动。 DataSync 支持的组件: Network File ...
  • SageMaker

    SageMaker 节点 综述 创建任务 任务样例 环境配置 SageMaker 节点 综述 Amazon SageMaker 是一个云机器学习平台。 提供了完整的基础设施,工具和工作流来帮助用户可以创建、训练和发布机器学习模型。 Amazon SageMaker Model Building Pipelines 是一个可以直接使用Sa...
  • Implementations

    1283 2024-05-25 《Apache JDO 3.2.1》
    JDO Reference Implementations Implementations To build and run your JDO application, you need a JDO implementation. This page lists commercial and non-commercial JDO implementat...
  • Customize Services

    Credentials Databases Directories Accounts All Configurations Next Step More Information The Customize Services step presents you with a set of tabs that let you review and...
  • Mixed-Iceberg

    1271 2024-06-26 《Apache Amoro 0.6.1》
    Compared with Iceberg format, Mixed-Iceberg format provides more features: Stronger primary key constraints that also apply to Spark OLAP performance that is production-ready fo...
  • 远程日志存储

    远程日志存储(Remote Logging) 开启远程日志存储 将任务日志写入 阿里云对象存储(OSS) 将任务日志写入 Amazon S3 将任务日志写入 Google Cloud Storage (GCS) 将任务日志写入 Azure Blob Storage (ABS) 注意事项 远程日志存储(Remote Logging) ...
  • High-Speed Ingest

    1262 2024-06-22 《Apache Accumulo 2.x》
    Pre-Splitting New Tables Multiple Ingest Clients Bulk Ingest Logical Time for Bulk Ingest MapReduce Ingest Accumulo is often used as part of a larger data processing and stor...
  • Hive

    1261 2024-06-29 《Apache Iceberg 1.5.2》
    Feature support Enabling Iceberg support in Hive Hive 4.0.0-beta-1 Hive 4.0.0-alpha-2 Hive 4.0.0-alpha-1 Hive 2.3.x, Hive 3.1.x Loading runtime jar Enabling support Hadoop con...
  • Customize Services

    Credentials Databases Directories Accounts All Configurations Next Step More Information The Customize Services step presents you with a set of tabs that let you review and...
  • Creating your first interoperable table

    Pre-requisites Steps Initialize a pyspark shell Create dataset Running sync Conclusion Next steps Using OneTable to sync your source tables in different target format invo...