-
Notifications
You must be signed in to change notification settings - Fork 972
[KYUUBI #3441] Change default Spark version to 3.3.1 #3507
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
@pan3793 , Tracking from here |
...connector-tpcds/src/test/scala/org/apache/kyuubi/spark/connector/tpcds/TPCDSQuerySuite.scala
Outdated
Show resolved
Hide resolved
|
FYI: Spark-3.3.1-rc2 is out. |
|
@cfmcgrady Thanks, will test |
|
The test failed consistently, could you please take a look? |
will check |
| run: >- | ||
| ./build/mvn ${MVN_OPT} clean install | ||
| -Pflink-provided,hive-provided | ||
| -Pspark-3.2 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It looks like that we currently support spark 3.2 only, We should specify the profile here to prevent some incompatibilities (on k8s w/ client mode)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
it's tricky, what's the specific issue here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In spark on k8s w/ client mode, task deserialization may fail due to inconsistent Jar packages on the Driver and Executor ends
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
oh sorry, i missed we hardcoded the spark image in code, how about changing it to get from ENV then we can override the in test?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We can get the compile Spark version by
SPARK_VERSION=$("$MVN" help:evaluate -Dexpression=spark.version $@ 2>/dev/null\
| grep -v "INFO"\
| grep -v "WARNING"\
| tail -n 1)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, I had the idea that we could pass the spark version of the current profile through the maven plugin so that the image version would be consistent with the spark version of the current profile, but I wanted to do this later for two reasons:
- This is to address the current todo and seems to be unrelated to the current pr
- The official spark image does not have a sub-version like Spark 3.3.1 https://hub.docker.com/r/apache/spark/tags, the Spark community is currently working on making the spark image the official docker image, maybe we can solve this todo after that?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
make sense, thanks
pan3793
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, it's in good shape now, let's wait for the next RC or GA version, thanks @yikf
|
Test for Spark 3.3.1 (RC4) |
|
Spark 3.3.1 (RC4) vote passed. |
|
Thanks all. |
Why are the changes needed?
Fix #3441;
This pr aims to two points as follow:
How was this patch tested?
Add some test cases that check the changes thoroughly including negative and positive cases if possible
Add screenshots for manual tests if appropriate
Run test locally before make a pull request