当前位置:网站首页>Apache seatunnel 2.1.0 deployment and stepping on the pit
Apache seatunnel 2.1.0 deployment and stepping on the pit
2022-04-23 13:42:00 【Ruo Xiaoyu】
brief introduction
SeaTunnel Original name Waterdrop, since 2021 year 10 month 12 Renamed SeaTunnel.
SeaTunnel It is a very easy to use ultra-high performance distributed data integration platform , Support real-time synchronization of massive data . It can synchronize tens of billions of data stably and efficiently every day , It's near 100 Used in the production of this company .
characteristic
- Easy to use , Flexible configuration , Low code development
- Real time streaming
- Offline multi-source data analysis
- High performance 、 Massive data processing capabilities
- Modular and plug-in mechanisms , extensible
- Supported by SQL Data processing and aggregation
- Support Spark Structured streaming media
- Support Spark 2.x
- Here we stepped on a pit , Because we tested spark The environment has been upgraded to 3.x edition , Now, SeaTunnel Only support 2.x, So we need to redeploy one 2.x Of spark
-
- Here we stepped on a pit , Because we tested spark The environment has been upgraded to 3.x edition , Now, SeaTunnel Only support 2.x, So we need to redeploy one 2.x Of spark
Workflow
install
Installation document
https://seatunnel.incubator.apache.org/docs/2.1.0/spark/installation
- Environmental preparation : install jdk and spark
- config/seatunnel-env.sh
- Download installation package
- https://www.apache.org/dyn/closer.lua/incubator/seatunnel/2.1.0/apache-seatunnel-incubating-2.1.0-bin.tar.gz
- Decompress and edit config/seatunnel-env.sh
- Specify the necessary environment configuration , for example SPARK_HOME(SPARK Download and unzip the directory )
1、 test jdbc-to-jdbc
- Create a new config/spark.batch.jdbc.to.jdbc.conf file
env {
# seatunnel defined streaming batch duration in seconds
spark.app.name = "SeaTunnel"
spark.executor.instances = 1
spark.executor.cores = 1
spark.executor.memory = "1g"
}
source {
jdbc {
driver = "com.mysql.jdbc.Driver"
url = "jdbc:mysql://0.0.0.0:3306/database?useUnicode=true&characterEncoding=utf8&useSSL=false"
table = "table_name"
result_table_name = "result_table_name"
user = "root"
password = "password"
}
}
transform {
# split data by specific delimiter
# you can also use other filter plugins, such as sql
# sql {
# sql = "select * from accesslog where request_time > 1000"
# }
# If you would like to get more information about how to configure seatunnel and see full list of filter plugins,
# please go to https://seatunnel.apache.org/docs/spark/configuration/transform-plugins/Sql
}
sink {
# choose stdout output plugin to output data to console
# Console {}
jdbc {
# Configuration here driver Parameters , Otherwise, the data exchange will not succeed
driver = "com.mysql.jdbc.Driver",
saveMode = "update",
url = "jdbc:mysql://ip:3306/database?useUnicode=true&characterEncoding=utf8&useSSL=false",
user = "userName",
password = "***********",
dbTable = "tableName",
customUpdateStmt = "INSERT INTO table (column1, column2, created, modified, yn) values(?, ?, now(), now(), 1) ON DUPLICATE KEY UPDATE column1 = IFNULL(VALUES (column1), column1), column2 = IFNULL(VALUES (column2), column2)"
}
}
yarn Start command
./bin/start-seatunnel-spark.sh --master 'yarn' --deploy-mode client --config ./config/spark.batch.jdbc.to.jdbc.conf
Step on the pit : Run times [driver] as non-empty , Locate and find sink It needs to be set in the configuration driver Parameters
ERROR Seatunnel:121 - Plugin[org.apache.seatunnel.spark.sink.Jdbc] contains invalid config, error: please specify [driver] as non-empty
版权声明
本文为[Ruo Xiaoyu]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/04/202204230602186365.html
边栏推荐
- Oracle database combines the query result sets of multiple columns into one row
- PG library checks the name
- UEFI learning 01-arm aarch64 compilation, armplatformpripeicore (SEC)
- [point cloud series] learning representations and generative models for 3D point clouds
- [point cloud series] so net: self organizing network for point cloud analysis
- Antd design form verification
- Oracle kills the executing SQL
- NPM err code 500 solution
- Software test system integration project management engineer full truth simulation question (including answer and analysis)
- Plato farm, a top-level metauniverse game, has made frequent positive moves recently
猜你喜欢
TERSUS笔记员工信息516-Mysql查询(2个字段的时间段唯一性判断)
[point cloud series] unsupervised multi task feature learning on point clouds
On the bug of JS regular test method
SAP UI5 应用开发教程之七十二 - SAP UI5 页面路由的动画效果设置
MySQL 8.0.11 download, install and connect tutorials using visualization tools
Special window function rank, deny_ rank, row_ number
顶级元宇宙游戏Plato Farm,近期动作不断利好频频
The interviewer dug a hole for me: how many concurrent TCP connections can a single server have?
[Video] Bayesian inference in linear regression and R language prediction of workers' wage data | data sharing
Double pointer instrument panel reading (I)
随机推荐
Oracle and MySQL batch query all table names and table name comments under users
[point cloud series] multi view neural human rendering (NHR)
Comparison and summary of applicable scenarios of Clickhouse and MySQL database
零拷贝技术
Oracle calculates the difference between two dates in seconds, minutes, hours and days
Information: 2021 / 9 / 29 10:01 - build completed with 1 error and 0 warnings in 11S 30ms error exception handling
Oracle job scheduled task usage details
[andorid] realize SPI communication between kernel and app through JNI
NPM err code 500 solution
Window analysis function last_ VALUE,FIRST_ VALUE,lag,lead
[tensorflow] sharing mechanism
SHA512 / 384 principle and C language implementation (with source code)
Oracle clear SQL cache
Storage scheme of video viewing records of users in station B
Processbuilder tool class
The difference between is and as in Oracle stored procedure
[point cloud series] pointfilter: point cloud filtering via encoder decoder modeling
面试官给我挖坑:URI中的 “//” 有什么用?
Solve the problem that Oracle needs to set IP every time in the virtual machine
You and the 42W bonus pool are one short of the "Changsha bank Cup" Tencent yunqi innovation competition!