diff --git a/debezium-server-iceberg-sink/pom.xml b/debezium-server-iceberg-sink/pom.xml
index f83f70cb..cfdcf96e 100644
--- a/debezium-server-iceberg-sink/pom.xml
+++ b/debezium-server-iceberg-sink/pom.xml
@@ -93,11 +93,21 @@
iceberg-gcp
${version.iceberg}
+
+ org.apache.iceberg
+ iceberg-gcp-bundle
+ ${version.iceberg}
+
org.apache.iceberg
iceberg-aws
${version.iceberg}
+
+ org.apache.iceberg
+ iceberg-aws-bundle
+ ${version.iceberg}
+
org.apache.iceberg
iceberg-dell
@@ -113,15 +123,12 @@
iceberg-azure
${version.iceberg}
-
-
- com.google.cloud
- google-cloud-storage
-
- com.google.cloud
- google-cloud-nio
+ org.apache.iceberg
+ iceberg-azure-bundle
+ ${version.iceberg}
+
com.google.cloud.bigdataoss
gcs-connector
@@ -132,37 +139,6 @@
gcsio
${version.googlebigdataoss}
-
-
- software.amazon.awssdk
- core
- ${version.awssdk}
- pom
-
-
- software.amazon.awssdk
- glue
-
-
- software.amazon.awssdk
- s3
-
-
- software.amazon.awssdk
- dynamodb
-
-
- software.amazon.awssdk
- kms
-
-
- software.amazon.awssdk
- sts
-
-
- software.amazon.awssdk
- url-connection-client
-
org.apache.hadoop
hadoop-aws
diff --git a/debezium-server-iceberg-sink/src/test/java/io/debezium/server/iceberg/testresources/BaseSparkTest.java b/debezium-server-iceberg-sink/src/test/java/io/debezium/server/iceberg/testresources/BaseSparkTest.java
index f53131eb..f81fd743 100644
--- a/debezium-server-iceberg-sink/src/test/java/io/debezium/server/iceberg/testresources/BaseSparkTest.java
+++ b/debezium-server-iceberg-sink/src/test/java/io/debezium/server/iceberg/testresources/BaseSparkTest.java
@@ -47,35 +47,18 @@ static void setup() {
.set("spark.eventLog.enabled", "false")
// enable iceberg SQL Extensions and Catalog
.set("spark.sql.extensions", "org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions")
+ .set("spark.sql.warehouse.dir", S3_BUCKET)
// hadoop catalog
.set("spark.sql.catalog.spark_catalog", "org.apache.iceberg.spark.SparkSessionCatalog")
.set("spark.sql.catalog.spark_catalog.type", "hadoop")
.set("spark.sql.catalog.spark_catalog.warehouse", S3_BUCKET)
.set("spark.sql.catalog.spark_catalog.default-namespaces", CATALOG_TABLE_NAMESPACE)
.set("spark.sql.catalog.spark_catalog.io-impl", "org.apache.iceberg.aws.s3.S3FileIO")
- .set("spark.sql.warehouse.dir", S3_BUCKET)
.set("spark.sql.catalog.spark_catalog.s3.endpoint", "http://localhost:" + S3Minio.getMappedPort().toString())
.set("spark.sql.catalog.spark_catalog.s3.path-style-access", "true")
.set("spark.sql.catalog.spark_catalog.s3.access-key-id", S3Minio.MINIO_ACCESS_KEY)
.set("spark.sql.catalog.spark_catalog.s3.secret-access-key", S3Minio.MINIO_SECRET_KEY)
.set("spark.sql.catalog.spark_catalog.client.region", TestConfigSource.S3_REGION)
- .set("spark.sql.catalog.spark_catalog.io-impl", "org.apache.iceberg.aws.s3.S3FileIO")
- .set("spark.sql.catalog.spark_catalog.warehouse", S3_BUCKET)
-// // JdbcCatalog catalog, add additional catalog
-// .set("spark.sql.defaultCatalog", ICEBERG_CATALOG_NAME)
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME, "org.apache.iceberg.spark.SparkCatalog")
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".warehouse", S3_BUCKET)
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".cache-enabled", "false")
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".catalog-impl", JdbcCatalog.class.getName())
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".default-namespaces", CATALOG_TABLE_NAMESPACE)
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".uri", JdbcCatalogDB.container.getJdbcUrl())
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".jdbc.user", JdbcCatalogDB.container.getUsername())
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".jdbc.password", JdbcCatalogDB.container.getPassword())
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".io-impl", "org.apache.iceberg.aws.s3.S3FileIO")
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".s3.endpoint", "http://localhost:" + S3Minio.getMappedPort().toString())
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".s3.path-style-access", "true")
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".s3.access-key-id", S3Minio.MINIO_ACCESS_KEY)
-// .set("spark.sql.catalog." + ICEBERG_CATALOG_NAME + ".s3.secret-access-key", S3Minio.MINIO_SECRET_KEY)
;
BaseSparkTest.spark = SparkSession
diff --git a/pom.xml b/pom.xml
index 447feb29..4e348ff4 100644
--- a/pom.xml
+++ b/pom.xml
@@ -39,10 +39,6 @@
${version.spark.major}.1
3.3.6
3.1.3
-
- 2.24.5
-
- 26.28.0
2.2.20
1.19.6
@@ -74,14 +70,6 @@
slf4j-jboss-logmanager
[1.2.0.Final,)
-
-
- software.amazon.awssdk
- bom
- ${version.awssdk}
- pom
- import
-
io.quarkus
quarkus-bom
@@ -89,14 +77,6 @@
pom
import
-
-
- com.google.cloud
- libraries-bom
- ${version.googlelibraries}
- pom
- import
-
mysql
@@ -111,27 +91,13 @@
pom
import
-
-
- io.debezium
- debezium-server-batch
- ${project.version}
-
org.codehaus.groovy
- groovy
- ${version.groovy}
-
-
- org.codehaus.groovy
- groovy-json
- ${version.groovy}
-
-
- org.codehaus.groovy
- groovy-jsr223
+ groovy-bom
${version.groovy}
+ pom
+ import