Devlive 开源社区 本次搜索耗时 0.027 秒,为您找到 2342 个相关结果.
  • SSO 集成

    背景介绍 SSO 登录工作流 如何启用SSO登录 注意事项 背景介绍 作为企业实践,在应用程序中应用单点登录 (SSO) 是很常见的,这样可以通过集中且安全的方式管理用户凭证。 基于 Streampark 使用 Apache Shiro 进行身份验证和授权的事实,我们将使用 Pac4j 框架来实现单点登录 (SSO) 支持功能。 Pac4j 是...
  • Select Version

    Choosing Stack Choosing Version Choosing Repositories redhat7 sles12 ubuntu14 ubuntu16 ubuntu18 debian9 Advanced Options More Information In this Step, you will select ...
  • Nessie

    3980 2025-03-11 《Apache Iceberg 1.8.1》
    Iceberg Nessie Integration Iceberg provides integration with Nessie through the iceberg-nessie module. This section describes how to use Iceberg with Nessie. Nessie provides seve...
  • 9.1 LaTeX or HTML output

    3978 2024-05-24 《R Markdown Cookbook》
    LaTeX and HTML are two commonly used output formats. The function knitr::is_latex_output() tells you if the output format is LaTeX (including Pandoc output formats latex and bea...
  • Ubuntu 18

    Steps Next Step More Information On a server host that has Internet access, use a command line editor to perform the following: Steps Log in to your host as root . Download ...
  • deployment

    Step 1: Prepare the environment Step 2: Download SeaTunnel Step 3: Install connectors plugin What’s More Step 1: Prepare the environment Before you getting start the local ru...
  • Docker Demo

    Building interoperable tables using OneTable Pre-requisites Setting up Docker cluster Accessing Services Trino Jupyter Notebook Building interoperable tables using OneTable...
  • Sqoop

    SQOOP 节点 综述 创建任务 任务参数 任务样例 配置 Sqoop 环境 配置 Sqoop 任务节点 查看运行结果 SQOOP 节点 综述 SQOOP 任务类型,用于执行 SQOOP 程序。对于 SQOOP 节点,worker 会通过执行 sqoop 命令来执行 SQOOP 任务。 创建任务 点击项目管理 -> 项目名称 -...
  • Ubuntu 16

    Steps Next Step More Information On a server host that has Internet access, use a command line editor to perform the following: Steps Log in to your host as root . Download ...
  • AWS S3

    3964 2024-07-01 《Apache Hudi 0.15.0》
    AWS configs AWS Credentials AWS Libs AWS S3 Versioned Bucket In this page, we explain how to get your Hudi spark job to store into AWS S3. AWS configs There are two configur...