Devlive 开源社区 本次搜索耗时 0.402 秒,为您找到 445 个相关结果.
  • Implement SpotBugs plugin

    1316 2024-06-13 《SpotBugs 4.8.5》
    Implement SpotBugs plugin Create Maven project Write java code to represent bug to find Write test case to ensure your detector can find bug Write java code to avoid false-posit...
  • Apache Iceberg

    Support Iceberg Version Support Those Engines Key features Description Supported DataSource Info Database Dependency Data Type Mapping Source Options Task Example Simple: ...
  • SQL Functions

    String Functions ASCII CHAR_LENGTH / LENGTH OCTET_LENGTH CONCAT CONCAT_WS HEXTORAW LOWER / LCASE UPPER / UCASE LEFT RIGHT LOCATE / INSTR / POSITION LPAD RPAD LTRIM RT...
  • Flink on K8s

    Environments requirement Preparation for integration configuration for connecting Kubernetes configuration for coKubernetes RBAC Configuration for remote Docker service Job s...
  • Falcon 插件

    Nightingale 实现了 Open-Falcon 的 HTTP 数据接收接口,所以,Open-Falcon 社区很多采集插件也是可以直接使用的。但是 Falcon-Agent 不行,因为 Falcon-Agent 推送监控数据给服务端,走的是 RPC 接口。 如果你发现某个 Falcon-Plugin 是用 CRON 的方式驱动的,推送数据走的是 ...
  • Function

    1307 2024-06-14 《Lodash 3.10.1》
    _.after(n, func) Arguments Returns Example _.ary(func, [n=func.length]) Arguments Returns Example _.before(n, func) Arguments Returns Example _.bind(func, thisArg, [par...
  • List

    1306 2024-06-05 《Ramda 0.30.1》
    all any append concat drop zipWith zip xprod uniq filter find flatten head indexOf join lastIndexOf map nth pluck prepend range reduce reduceRight reject re...
  • Batch Writes

    1304 2024-06-30 《Apache Hudi 0.15.0》
    Spark DataSource API The hudi-spark module offers the DataSource API to write a Spark DataFrame into a Hudi table. There are a number of options available: HoodieWriteConfig :...
  • LocalFile

    Support Those Engines Key features Description Options path [string] file_format_type [string] read_columns [list] delimiter/field_delimiter [string] parse_partition_from_pa...
  • Bootstrapping

    1298 2024-06-30 《Apache Hudi 0.15.0》
    Approaches Use Hudi for new partitions alone Convert existing table to Hudi Using Hudi Streamer Using Spark Datasource Writer Using Spark SQL CALL Procedure Using Hudi CLI C...