Skip to content

Commit fa42049

Browse files
author
litiliu
committed
update to 3.4.1 & update to aws sdkV2
1 parent 35f41a7 commit fa42049

File tree

88 files changed

+423
-248
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

88 files changed

+423
-248
lines changed

docs/en/connector-v2/sink/Hive.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Write data to Hive.
1212

1313
In order to use this connector, You must ensure your spark/flink cluster already integrated hive. The tested hive version is 2.3.9 and 3.1.3 .
1414

15-
If you use SeaTunnel Engine, You need put seatunnel-hadoop3-3.1.4-uber.jar and hive-exec-3.1.3.jar and libfb303-0.9.3.jar in $SEATUNNEL_HOME/lib/ dir.
15+
If you use SeaTunnel Engine, You need put seatunnel-hadoop3-3.4.1-uber.jar and hive-exec-3.1.3.jar and libfb303-0.9.3.jar in $SEATUNNEL_HOME/lib/ dir.
1616
:::
1717

1818
## Key features
@@ -189,7 +189,7 @@ sink {
189189
metastore_uri = "thrift://ctyun7:9083"
190190
hive.hadoop.conf = {
191191
bucket = "s3a://mybucket"
192-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
192+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
193193
}
194194
}
195195
```
@@ -344,7 +344,7 @@ sink {
344344
hive.hadoop.conf-path = "/home/ec2-user/hadoop-conf"
345345
hive.hadoop.conf = {
346346
bucket="s3://ws-package"
347-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
347+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
348348
}
349349
}
350350
}

docs/en/connector-v2/sink/ObsFile.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,12 +34,11 @@ By default, we use 2PC commit to ensure `exactly-once`
3434

3535
Output data to huawei cloud obs file system.
3636

37-
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
37+
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 3.4.
3838

3939
If you use SeaTunnel Engine, It automatically integrated the hadoop jar when you download and install SeaTunnel Engine. You can check the jar package under ${SEATUNNEL_HOME}/lib to confirm this.
4040

4141
We made some trade-offs in order to support more file types, so we used the HDFS protocol for internal access to OBS and this connector need some hadoop dependencies.
42-
It only supports hadoop version **2.9.X+**.
4342

4443
## Required Jar List
4544

docs/en/connector-v2/sink/OssFile.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@ import ChangeLog from '../changelog/connector-file-oss.md';
1515
### For Spark/Flink Engine
1616

1717
1. You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
18-
2. You must ensure `hadoop-aliyun-xx.jar`, `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` in `${SEATUNNEL_HOME}/plugins/` dir and the version of `hadoop-aliyun` jar need equals your hadoop version which used in spark/flink and `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` version needs to be the version corresponding to the `hadoop-aliyun` version. Eg: `hadoop-aliyun-3.1.4.jar` dependency `aliyun-sdk-oss-3.4.1.jar` and `jdom-1.1.jar`.
18+
2. You must ensure `hadoop-aliyun-xx.jar`, `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` in `${SEATUNNEL_HOME}/plugins/` dir and the version of `hadoop-aliyun` jar need equals your hadoop version which used in spark/flink and `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` version needs to be the version corresponding to the `hadoop-aliyun` version. Eg: `hadoop-aliyun-3.4.1.jar` dependency `aliyun-sdk-oss-3.4.1.jar` and `jdom-1.1.jar`.
1919

2020
### For SeaTunnel Zeta Engine
2121

22-
1. You must ensure `seatunnel-hadoop3-3.1.4-uber.jar`, `aliyun-sdk-oss-3.4.1.jar`, `hadoop-aliyun-3.1.4.jar` and `jdom-1.1.jar` in `${SEATUNNEL_HOME}/lib/` dir.
22+
1. You must ensure `seatunnel-hadoop3-3.4.1-uber.jar`, `aliyun-sdk-oss-3.4.1.jar`, `hadoop-aliyun-3.4.1.jar` and `jdom-1.1.jar` in `${SEATUNNEL_HOME}/lib/` dir.
2323

2424
## Key features
2525

docs/en/connector-v2/sink/OssJindoFile.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,12 +13,11 @@ Output data to oss file system using jindo api.
1313
You need to download [jindosdk-4.6.1.tar.gz](https://jindodata-binary.oss-cn-shanghai.aliyuncs.com/release/4.6.1/jindosdk-4.6.1.tar.gz)
1414
and then unzip it, copy jindo-sdk-4.6.1.jar and jindo-core-4.6.1.jar from lib to ${SEATUNNEL_HOME}/lib.
1515

16-
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
16+
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 3.4.
1717

1818
If you use SeaTunnel Engine, It automatically integrated the hadoop jar when you download and install SeaTunnel Engine. You can check the jar package under ${SEATUNNEL_HOME}/lib to confirm this.
1919

2020
We made some trade-offs in order to support more file types, so we used the HDFS protocol for internal access to OSS and this connector need some hadoop dependencies.
21-
It only supports hadoop version **2.9.X+**.
2221

2322
:::
2423

docs/en/connector-v2/sink/S3File.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ Output data to aws s3 file system.
4848
> If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
4949
>
5050
> If you use SeaTunnel Engine, It automatically integrated the hadoop jar when you download and install SeaTunnel Engine. You can check the jar package under `${SEATUNNEL_HOME}/lib` to confirm this.
51-
> To use this connector you need put `hadoop-aws-3.1.4.jar` and `aws-java-sdk-bundle-1.12.692.jar` in `${SEATUNNEL_HOME}/lib` dir.
51+
> To use this connector, place the seatunnel-hadoop-aws.jar (which includes the bundled bundle.jar with AWS SDK v2) in the ${SEATUNNEL_HOME}/lib directory.
5252
5353
## Data Type Mapping
5454

@@ -106,7 +106,7 @@ If write to `csv`, `text` file type, All column will be string.
106106
| tmp_path | string | no | /tmp/seatunnel | The result file will write to a tmp path first and then use `mv` to submit tmp dir to target dir. Need a S3 dir. |
107107
| bucket | string | yes | - | |
108108
| fs.s3a.endpoint | string | yes | - | |
109-
| fs.s3a.aws.credentials.provider | string | yes | com.amazonaws.auth.InstanceProfileCredentialsProvider | The way to authenticate s3a. We only support `org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider` and `com.amazonaws.auth.InstanceProfileCredentialsProvider` now. |
109+
| fs.s3a.aws.credentials.provider | string | yes | software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider | The way to authenticate s3a. We only support `org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider` and `software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider` now. |
110110
| access_key | string | no | - | Only used when fs.s3a.aws.credentials.provider = org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider |
111111
| secret_key | string | no | - | Only used when fs.s3a.aws.credentials.provider = org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider |
112112
| custom_filename | boolean | no | false | Whether you need custom the filename |
@@ -377,7 +377,7 @@ sink {
377377
tmp_path = "/tmp/seatunnel"
378378
path="/seatunnel/text"
379379
fs.s3a.endpoint="s3.cn-north-1.amazonaws.com.cn"
380-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
380+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
381381
file_format_type = "text"
382382
field_delimiter = "\t"
383383
row_delimiter = "\n"
@@ -401,7 +401,7 @@ sink {
401401
```
402402

403403
For text file format with `have_partition` and `custom_filename` and `sink_columns`
404-
and `com.amazonaws.auth.InstanceProfileCredentialsProvider`
404+
and `software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider`
405405

406406
```hocon
407407
@@ -410,7 +410,7 @@ and `com.amazonaws.auth.InstanceProfileCredentialsProvider`
410410
tmp_path = "/tmp/seatunnel"
411411
path="/seatunnel/text"
412412
fs.s3a.endpoint="s3.cn-north-1.amazonaws.com.cn"
413-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
413+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
414414
file_format_type = "text"
415415
field_delimiter = "\t"
416416
row_delimiter = "\n"

docs/en/connector-v2/source/Hive.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Read data from Hive.
1212

1313
In order to use this connector, You must ensure your spark/flink cluster already integrated hive. The tested hive version is 2.3.9 and 3.1.3 .
1414

15-
If you use SeaTunnel Engine, You need put seatunnel-hadoop3-3.1.4-uber.jar and hive-exec-3.1.3.jar and libfb303-0.9.3.jar in $SEATUNNEL_HOME/lib/ dir.
15+
If you use SeaTunnel Engine, You need put seatunnel-hadoop3-3.4.1-uber.jar and hive-exec-3.1.3.jar and libfb303-0.9.3.jar in $SEATUNNEL_HOME/lib/ dir.
1616
:::
1717

1818
## Key features
@@ -295,7 +295,7 @@ source {
295295
hive.hadoop.conf-path = "/home/ec2-user/hadoop-conf"
296296
hive.hadoop.conf = {
297297
bucket="s3://ws-package"
298-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
298+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
299299
}
300300
read_columns = ["pk_id", "name", "score"]
301301
}
@@ -308,7 +308,7 @@ sink {
308308
hive.hadoop.conf-path = "/home/ec2-user/hadoop-conf"
309309
hive.hadoop.conf = {
310310
bucket="s3://ws-package"
311-
fs.s3a.aws.credentials.provider="com.amazonaws.auth.InstanceProfileCredentialsProvider"
311+
fs.s3a.aws.credentials.provider="software.amazon.awssdk.auth.credentials.InstanceProfileCredentialsProvider"
312312
}
313313
}
314314
}

docs/en/connector-v2/source/Jdbc.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -269,7 +269,7 @@ there are some reference value for params above.
269269
| Vertica | com.vertica.jdbc.Driver | jdbc:vertica://localhost:5433 | https://repo1.maven.org/maven2/com/vertica/jdbc/vertica-jdbc/12.0.3-0/vertica-jdbc-12.0.3-0.jar |
270270
| Kingbase | com.kingbase8.Driver | jdbc:kingbase8://localhost:54321/db_test | https://repo1.maven.org/maven2/cn/com/kingbase/kingbase8/8.6.0/kingbase8-8.6.0.jar |
271271
| OceanBase | com.oceanbase.jdbc.Driver | jdbc:oceanbase://localhost:2881 | https://repo1.maven.org/maven2/com/oceanbase/oceanbase-client/2.4.12/oceanbase-client-2.4.12.jar |
272-
| Hive | org.apache.hive.jdbc.HiveDriver | jdbc:hive2://localhost:10000 | https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/3.1.3/hive-jdbc-3.1.3-standalone.jar |
272+
| Hive | org.apache.hive.jdbc.HiveDriver | jdbc:hive2://localhost:10000 | https://repo1.maven.org/maven2/org/apache/hive/hive-jdbc/4.0.1/hive-jdbc-4.0.1-standalone.jar |
273273
| xugu | com.xugu.cloudjdbc.Driver | jdbc:xugu://localhost:5138 | https://repo1.maven.org/maven2/com/xugudb/xugu-jdbc/12.2.0/xugu-jdbc-12.2.0.jar |
274274
| InterSystems IRIS | com.intersystems.jdbc.IRISDriver | jdbc:IRIS://localhost:1972/%SYS | https://raw.githubusercontent.com/intersystems-community/iris-driver-distribution/main/JDBC/JDK18/intersystems-jdbc-3.8.4.jar |
275275
| opengauss | org.opengauss.Driver | jdbc:opengauss://localhost:5432/postgres | https://repo1.maven.org/maven2/org/opengauss/opengauss-jdbc/5.1.0-og/opengauss-jdbc-5.1.0-og.jar |

docs/en/connector-v2/source/ObsFile.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,12 +39,11 @@ import ChangeLog from '../changelog/connector-file-obs.md';
3939

4040
Read data from huawei cloud obs file system.
4141

42-
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
42+
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 3.4.
4343

4444
If you use SeaTunnel Engine, It automatically integrated the hadoop jar when you download and install SeaTunnel Engine. You can check the jar package under ${SEATUNNEL_HOME}/lib to confirm this.
4545

4646
We made some trade-offs in order to support more file types, so we used the HDFS protocol for internal access to OBS and this connector need some hadoop dependencies.
47-
It only supports hadoop version **2.9.X+**.
4847

4948
## Required Jar List
5049

docs/en/connector-v2/source/OssFile.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@ import ChangeLog from '../changelog/connector-file-oss.md';
1515
### For Spark/Flink Engine
1616

1717
1. You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
18-
2. You must ensure `hadoop-aliyun-xx.jar`, `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` in `${SEATUNNEL_HOME}/plugins/` dir and the version of `hadoop-aliyun` jar need equals your hadoop version which used in spark/flink and `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` version needs to be the version corresponding to the `hadoop-aliyun` version. Eg: `hadoop-aliyun-3.1.4.jar` dependency `aliyun-sdk-oss-3.4.1.jar` and `jdom-1.1.jar`.
18+
2. You must ensure `hadoop-aliyun-xx.jar`, `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` in `${SEATUNNEL_HOME}/plugins/` dir and the version of `hadoop-aliyun` jar need equals your hadoop version which used in spark/flink and `aliyun-sdk-oss-xx.jar` and `jdom-xx.jar` version needs to be the version corresponding to the `hadoop-aliyun` version. Eg: `hadoop-aliyun-3.4.1.jar` dependency `aliyun-sdk-oss-3.4.1.jar` and `jdom-1.1.jar`.
1919

2020
### For SeaTunnel Zeta Engine
2121

22-
1. You must ensure `seatunnel-hadoop3-3.1.4-uber.jar`, `aliyun-sdk-oss-3.4.1.jar`, `hadoop-aliyun-3.1.4.jar` and `jdom-1.1.jar` in `${SEATUNNEL_HOME}/lib/` dir.
22+
1. You must ensure `seatunnel-hadoop3-3.4.1-uber.jar`, `aliyun-sdk-oss-3.4.1.jar`, `hadoop-aliyun-3.4.1.jar` and `jdom-1.1.jar` in `${SEATUNNEL_HOME}/lib/` dir.
2323

2424
## Key features
2525

docs/en/connector-v2/source/OssJindoFile.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -44,12 +44,11 @@ Read data from aliyun oss file system using jindo api.
4444
You need to download [jindosdk-4.6.1.tar.gz](https://jindodata-binary.oss-cn-shanghai.aliyuncs.com/release/4.6.1/jindosdk-4.6.1.tar.gz)
4545
and then unzip it, copy jindo-sdk-4.6.1.jar and jindo-core-4.6.1.jar from lib to ${SEATUNNEL_HOME}/lib.
4646

47-
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 2.x.
47+
If you use spark/flink, In order to use this connector, You must ensure your spark/flink cluster already integrated hadoop. The tested hadoop version is 3.4.
4848

4949
If you use SeaTunnel Engine, It automatically integrated the hadoop jar when you download and install SeaTunnel Engine. You can check the jar package under ${SEATUNNEL_HOME}/lib to confirm this.
5050

5151
We made some trade-offs in order to support more file types, so we used the HDFS protocol for internal access to OSS and this connector need some hadoop dependencies.
52-
It only supports hadoop version **2.9.X+**.
5352

5453
:::
5554

0 commit comments

Comments
 (0)