quick-start-spark
Quick Start With Spark
Step 1: Deployment SeaTunnel And Connectorsâ
Before starting, make sure you have downloaded and deployed SeaTunnel as described in deployment
Step 2: Deployment And Config Sparkâ
Please download Spark first(required version >= 2.4.0). For more information you could see Getting Started: standalone
Configure SeaTunnel: Change the setting in config/seatunnel-env.sh
, it is base on the path your engine install at deployment.
Change SPARK_HOME
to the Spark deployment dir.
Step 3: Add Job Config File to define a jobâ
Edit config/seatunnel.streaming.conf.template
, which determines the way and logic of data input, processing, and output after seatunnel is started.
The following is an example of the configuration file, which is the same as the example application mentioned above.
env {
execution.parallelism = 1
job.mode = "BATCH"
}
source {
FakeSource {
result_table_name = "fake"
row.num = 16
schema = {
fields {
name = "string"
age = "int"
}
}
}
}
transform {
FieldMapper {
source_table_name = "fake"
result_table_name = "fake1"
field_mapper = {
age = age
name = new_name
}
}
}
sink {
Console {
source_table_name = "fake1"
}
}
More information about config please check config concept
Step 3: Run SeaTunnel Applicationâ
You could start the application by the following commands
spark 2.4.x
cd "apache-seatunnel-${version}"
./bin/start-seatunnel-spark-2-connector-v2.sh \
--master local[4] \
--deploy-mode client \
--config ./config/seatunnel.streaming.conf.template
spark3.x.x
cd "apache-seatunnel-${version}"
./bin/start-seatunnel-spark-3-connector-v2.sh \
--master local[4] \
--deploy-mode client \
--config ./config/seatunnel.streaming.conf.template
See The Output: When you run the command, you could see its output in your console. You can think this is a sign that the command ran successfully or not.
The SeaTunnel console will prints some logs as below:
fields : name, age
types : STRING, INT
row=1 : elWaB, 1984352560
row=2 : uAtnp, 762961563
row=3 : TQEIB, 2042675010
row=4 : DcFjo, 593971283
row=5 : SenEb, 2099913608
row=6 : DHjkg, 1928005856
row=7 : eScCM, 526029657
row=8 : sgOeE, 600878991
row=9 : gwdvw, 1951126920
row=10 : nSiKE, 488708928
row=11 : xubpl, 1420202810
row=12 : rHZqb, 331185742
row=13 : rciGD, 1112878259
row=14 : qLhdI, 1457046294
row=15 : ZTkRx, 1240668386
row=16 : SGZCr, 94186144
What's Moreâ
For now, you are already take a quick look about SeaTunnel with Spark, you could see connector to find all source and sink SeaTunnel supported. Or see SeaTunnel With Spark if you want to know more about SeaTunnel Run With Spark.
SeaTunnel have an own engine named Zeta
and Zeta
is the default engine of SeaTunnel. You can follow Quick Start to configure and run a data synchronization job.