Devlive 开源社区 本次搜索耗时 0.377 秒,为您找到 760 个相关结果.
  • Procedures

    2073 2025-03-11 《Apache Iceberg 1.8.1》
    Spark Procedures To use Iceberg in Spark, first configure Spark catalogs . Stored procedures are only available when using Iceberg SQL extensions in Spark 3. Usage Procedures c...
  • Streaming Reads

    2063 2024-06-30 《Apache Hudi 0.15.0》
    Spark Streaming Spark Streaming Structured Streaming reads are based on Hudi’s Incremental Query feature, therefore streaming read can return data for which commits and base fil...
  • Socket

    Support Those Engines Key features Description Sink Options Task Example Changelog 2.2.0-beta 2022-09-26 Socket sink connector Support Those Engines Spark Flink SeaTu...
  • Flink Writes

    2055 2025-03-11 《Apache Iceberg 1.8.1》
    Flink Writes Iceberg support batch and streaming writes With Apache Flink ‘s DataStream API and Table API. Writing with SQL Iceberg support both INSERT INTO and INSERT OVERWRIT...
  • Relation

    2048 2024-06-05 《Ramda 0.2.0》
    countBy difference differenceWith unionWith union gt gte intersection lt lte max min propEq sortBy countBy (a → String) → [a] → {*} Parameters fnThe function u...
  • 第一章:多线程入门

    学习目标 1. 什么是线程与多线程 1.1 线程的概念 1.2 多线程的概念 2. 为什么需要多线程编程 2.1 提高CPU利用率 2.2 提高程序响应性 2.3 更好的资源利用 2.4 简化复杂问题的处理 3. Java中实现多线程的两种基本方式 3.1 继承Thread类 3.2 实现Runnable接口 3.3 两种方式的比较 ...
  • Tailwind CSS 移动端适配实践

    本节将详细介绍如何使用 Tailwind CSS 进行移动端适配,包括响应式设计、触摸交互优化、性能优化等方面。 基础配置 视口配置 <!-- public/index.html --> <meta name = "viewport" content = "width=device-width, initial-scale=1.0, max...
  • Tailwind CSS 部署环境配置

    在将 Tailwind CSS 项目部署到生产环境时,需要做好各项配置和优化工作。本节将详细介绍如何配置和优化部署环境。 环境配置 Docker 配置 # Dockerfile # 构建阶段 FROM node : 16 - alpine as builder WORKDIR / app COPY package *. ...
  • Object

    2017 2024-06-05 《Ramda 0.2.2》
    clone values eqProps keys omit pick pickAll project prop props where keysIn path valuesIn clone {*} → {*} Parameters valueThe object or array to clone Return...
  • Kafka source compatible kafka-connect-json

    How To Use Kafka Sink Mysql Seatunnel connector kafka supports parsing data extracted through kafka connect source, especially data extracted from kafka connect jdbc and kafka c...