Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
21 changes: 11 additions & 10 deletions dev/deps/spark-deps-hadoop-3-hive-2.3
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ aliyun-java-sdk-core/4.5.10//aliyun-java-sdk-core-4.5.10.jar
aliyun-java-sdk-kms/2.11.0//aliyun-java-sdk-kms-2.11.0.jar
aliyun-java-sdk-ram/3.1.0//aliyun-java-sdk-ram-3.1.0.jar
aliyun-sdk-oss/3.13.2//aliyun-sdk-oss-3.13.2.jar
analyticsaccelerator-s3/1.2.1//analyticsaccelerator-s3-1.2.1.jar
Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this won't be present at the official release artifacts, so we don't need to change the LICENSE/NOTICE files

Copy link
Member

@dongjoon-hyun dongjoon-hyun Aug 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What do you mean? If we don't want to change LICENSE/NOTICE, we need to exclude this explicitly.

Copy link
Member Author

@pan3793 pan3793 Aug 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

My understanding is: the LICENSE/NOTICE should match the content of the artifact.

For the source release tarball, in addition to Spark code itself, the LICENSE/NOTICE only reflects the source code we included from outside of the Spark project.

For the binary release tarball, the LICENSE-binary/NOTICE-binary should only reflect the included Spark and third-party libs, so LICENSE/NOTICE of testing deps and other optional libs is not necessary.

I'm not an expert in this area, the above is what I learned from several incubating projects.

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

$ dev/make-distribution.sh -Pyarn -Pkubernetes -Phadoop-3 -Phive -Phive-thriftserver
...
$ ls dist/jars | grep analyticsaccelerator
<no ouptut>

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Now I got your meaning. So, you mean we found a bug of dev/test-dependencies.sh, right? Could you file a bug JIRA issue for this independently?

Copy link
Member Author

@pan3793 pan3793 Aug 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

dev/test-dependencies.sh also gathers deps from optional modules, for example, the official release tarball does not include the hadoop-cloud module, thus it won't pull those transitive deps.

Copy link
Member

@dongjoon-hyun dongjoon-hyun Aug 28, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Got it~ It makes sense.

antlr-runtime/3.5.2//antlr-runtime-3.5.2.jar
antlr4-runtime/4.13.1//antlr4-runtime-4.13.1.jar
aopalliance-repackaged/3.0.6//aopalliance-repackaged-3.0.6.jar
Expand Down Expand Up @@ -70,16 +71,16 @@ gcs-connector/hadoop3-2.2.28/shaded/gcs-connector-hadoop3-2.2.28-shaded.jar
gmetric4j/1.0.10//gmetric4j-1.0.10.jar
gson/2.11.0//gson-2.11.0.jar
guava/33.4.0-jre//guava-33.4.0-jre.jar
hadoop-aliyun/3.4.1//hadoop-aliyun-3.4.1.jar
hadoop-annotations/3.4.1//hadoop-annotations-3.4.1.jar
hadoop-aws/3.4.1//hadoop-aws-3.4.1.jar
hadoop-azure-datalake/3.4.1//hadoop-azure-datalake-3.4.1.jar
hadoop-azure/3.4.1//hadoop-azure-3.4.1.jar
hadoop-client-api/3.4.1//hadoop-client-api-3.4.1.jar
hadoop-client-runtime/3.4.1//hadoop-client-runtime-3.4.1.jar
hadoop-cloud-storage/3.4.1//hadoop-cloud-storage-3.4.1.jar
hadoop-huaweicloud/3.4.1//hadoop-huaweicloud-3.4.1.jar
hadoop-shaded-guava/1.3.0//hadoop-shaded-guava-1.3.0.jar
hadoop-aliyun/3.4.2//hadoop-aliyun-3.4.2.jar
hadoop-annotations/3.4.2//hadoop-annotations-3.4.2.jar
hadoop-aws/3.4.2//hadoop-aws-3.4.2.jar
hadoop-azure-datalake/3.4.2//hadoop-azure-datalake-3.4.2.jar
hadoop-azure/3.4.2//hadoop-azure-3.4.2.jar
hadoop-client-api/3.4.2//hadoop-client-api-3.4.2.jar
hadoop-client-runtime/3.4.2//hadoop-client-runtime-3.4.2.jar
hadoop-cloud-storage/3.4.2//hadoop-cloud-storage-3.4.2.jar
hadoop-huaweicloud/3.4.2//hadoop-huaweicloud-3.4.2.jar
hadoop-shaded-guava/1.4.0//hadoop-shaded-guava-1.4.0.jar
hive-beeline/2.3.10//hive-beeline-2.3.10.jar
hive-cli/2.3.10//hive-cli-2.3.10.jar
hive-common/2.3.10//hive-common-2.3.10.jar
Expand Down
2 changes: 1 addition & 1 deletion pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -126,7 +126,7 @@
<slf4j.version>2.0.17</slf4j.version>
<log4j.version>2.25.1</log4j.version>
<!-- make sure to update IsolatedClientLoader whenever this version is changed -->
<hadoop.version>3.4.1</hadoop.version>
<hadoop.version>3.4.2</hadoop.version>
<!-- SPARK-41247: When updating `protobuf.version`, also need to update `protoVersion` in `SparkBuild.scala` -->
<protobuf.version>4.29.3</protobuf.version>
<protoc-jar-maven-plugin.version>3.11.4</protoc-jar-maven-plugin.version>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ private[hive] object IsolatedClientLoader extends Logging {
case e: RuntimeException if e.getMessage.contains("hadoop") =>
// If the error message contains hadoop, it is probably because the hadoop
// version cannot be resolved.
val fallbackVersion = "3.4.1"
val fallbackVersion = "3.4.2"
logWarning(log"Failed to resolve Hadoop artifacts for the version " +
log"${MDC(HADOOP_VERSION, hadoopVersion)}. We will change the hadoop version from " +
log"${MDC(HADOOP_VERSION, hadoopVersion)} to " +
Expand Down