-
Notifications
You must be signed in to change notification settings - Fork 7k
[spark] Validation of the Ray-on-Spark-on-YARN mode to enable it to run. #58335
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: Cai Zhanqi <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request adds support for running Ray on Spark with a YARN master by updating the validation logic. The change is straightforward and correct. I've suggested a minor improvement to the error message to include the detected Spark master, which will enhance debuggability for users encountering configuration issues.
| raise RuntimeError( | ||
| "Ray on Spark only supports spark cluster in standalone mode, " | ||
| "local-cluster mode or spark local mode." | ||
| "local-cluster mode, spark on yarn mode or spark local mode." | ||
| ) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
To improve debuggability, consider including the detected spark_master value in the error message. This will help users quickly identify misconfigurations when they are using an unsupported Spark master.
| raise RuntimeError( | |
| "Ray on Spark only supports spark cluster in standalone mode, " | |
| "local-cluster mode or spark local mode." | |
| "local-cluster mode, spark on yarn mode or spark local mode." | |
| ) | |
| raise RuntimeError( | |
| "Ray on Spark only supports spark cluster in standalone mode, " | |
| "local-cluster mode, spark on yarn mode or spark local mode. " | |
| f"Detected spark master is '{spark_master}'." | |
| ) |
|
@CaiZhanqi have you already tested running Ray on Spark-on-YARN mode? |
|
@edoakes yes, have tested Spark 3.2 by removing Spark version check, and Apache Spark 3.5 as well. |
…un. (ray-project#58335) ## Description > Add spark master model validation to let Ray run on Spark-On-YARN mode. ## Why need this? > If we directly run Ray on a YARN cluster, we need to do more tests and integration, and also need to setup related tools and environments. If support ray-on-spark-on-yarn and we already have Spark envs setup, we don't need to do other things, can use Spark and let the user run pyspark. Signed-off-by: Cai Zhanqi <[email protected]> Co-authored-by: Cai Zhanqi <[email protected]>
…un. (ray-project#58335) ## Description > Add spark master model validation to let Ray run on Spark-On-YARN mode. ## Why need this? > If we directly run Ray on a YARN cluster, we need to do more tests and integration, and also need to setup related tools and environments. If support ray-on-spark-on-yarn and we already have Spark envs setup, we don't need to do other things, can use Spark and let the user run pyspark. Signed-off-by: Cai Zhanqi <[email protected]> Co-authored-by: Cai Zhanqi <[email protected]>
…un. (ray-project#58335) ## Description > Add spark master model validation to let Ray run on Spark-On-YARN mode. ## Why need this? > If we directly run Ray on a YARN cluster, we need to do more tests and integration, and also need to setup related tools and environments. If support ray-on-spark-on-yarn and we already have Spark envs setup, we don't need to do other things, can use Spark and let the user run pyspark. Signed-off-by: Cai Zhanqi <[email protected]> Co-authored-by: Cai Zhanqi <[email protected]> Signed-off-by: Aydin Abiar <[email protected]>
…un. (ray-project#58335) ## Description > Add spark master model validation to let Ray run on Spark-On-YARN mode. ## Why need this? > If we directly run Ray on a YARN cluster, we need to do more tests and integration, and also need to setup related tools and environments. If support ray-on-spark-on-yarn and we already have Spark envs setup, we don't need to do other things, can use Spark and let the user run pyspark. Signed-off-by: Cai Zhanqi <[email protected]> Co-authored-by: Cai Zhanqi <[email protected]>
Description
Why need this?