Skip to main content
Version: 2.3.1

Hudi

Hudi source connector

Description

Used to read data from Hudi. Currently, only supports hudi cow table and Snapshot Query with Batch Mode.

In order to use this connector, You must ensure your spark/flink cluster already integrated hive. The tested hive version is 2.3.9.

Key features

Currently, only supports hudi cow table and Snapshot Query with Batch Mode

Options

nametyperequireddefault value
table.pathstringyes-
table.typestringyes-
conf.filesstringyes-
use.kerberosbooleannofalse
kerberos.principalstringyes when use.kerberos = true-
kerberos.principal.filestringyes when use.kerberos = true-
common-optionsconfigno-

table.path [string]

table.path The hdfs root path of hudi table,such as 'hdfs://nameserivce/data/hudi/hudi_table/'.

table.type [string]

table.type The type of hudi table. Now we only support 'cow', 'mor' is not support yet.

conf.files [string]

conf.files The environment conf file path list(local path), which used to init hdfs client to read hudi table file. The example is '/home/test/hdfs-site.xml;/home/test/core-site.xml;/home/test/yarn-site.xml'.

use.kerberos [boolean]

use.kerberos Whether to enable Kerberos, default is false.

kerberos.principal [string]

kerberos.principal When use kerberos, we should set kerberos princal such as 'test_user@xxx'.

kerberos.principal.file [string]

kerberos.principal.file When use kerberos, we should set kerberos princal file such as '/home/test/test_user.keytab'.

common options

Source plugin common parameters, please refer to Source Common Options for details.

Examples

source {

Hudi {
table.path = "hdfs://nameserivce/data/hudi/hudi_table/"
table.type = "cow"
conf.files = "/home/test/hdfs-site.xml;/home/test/core-site.xml;/home/test/yarn-site.xml"
use.kerberos = true
kerberos.principal = "test_user@xxx"
kerberos.principal.file = "/home/test/test_user.keytab"
}

}

Changelog

2.2.0-beta 2022-09-26

  • Add Hudi Source Connector