From 3c052e4a068355cddb930a9054301fb8ae9a50df Mon Sep 17 00:00:00 2001 From: HyukjinKwon Date: Sun, 22 Sep 2019 18:02:36 +0900 Subject: [PATCH 1/3] Run the tests of Cogrouped pandas UDF --- dev/sparktestsupport/modules.py | 1 + 1 file changed, 1 insertion(+) diff --git a/dev/sparktestsupport/modules.py b/dev/sparktestsupport/modules.py index d4e6b6749010..c7ea065b28ed 100644 --- a/dev/sparktestsupport/modules.py +++ b/dev/sparktestsupport/modules.py @@ -372,6 +372,7 @@ def __hash__(self): "pyspark.sql.tests.test_functions", "pyspark.sql.tests.test_group", "pyspark.sql.tests.test_pandas_udf", + "pyspark.sql.tests.test_pandas_udf_cogrouped_map", "pyspark.sql.tests.test_pandas_udf_grouped_agg", "pyspark.sql.tests.test_pandas_udf_grouped_map", "pyspark.sql.tests.test_pandas_udf_scalar", From 7d9835dc9cbba6a57c3cd190db62a7da0231671b Mon Sep 17 00:00:00 2001 From: HyukjinKwon Date: Sun, 22 Sep 2019 18:11:26 +0900 Subject: [PATCH 2/3] Skip Scala tests for now --- dev/run-tests.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/dev/run-tests.py b/dev/run-tests.py index ea515708124d..c93423e36c92 100755 --- a/dev/run-tests.py +++ b/dev/run-tests.py @@ -637,7 +637,7 @@ def main(): build_spark_assembly_sbt(hadoop_version, should_run_java_style_checks) # run the test suites - run_scala_tests(build_tool, hadoop_version, test_modules, excluded_tags) + # run_scala_tests(build_tool, hadoop_version, test_modules, excluded_tags) modules_with_python_tests = [m for m in test_modules if m.python_test_goals] if modules_with_python_tests: From f6cd95a3d1982ee2558cf49b0491a9e9ace47598 Mon Sep 17 00:00:00 2001 From: HyukjinKwon Date: Sun, 22 Sep 2019 18:42:35 +0900 Subject: [PATCH 3/3] Revert "Skip Scala tests for now" This reverts commit 7d9835dc9cbba6a57c3cd190db62a7da0231671b. --- dev/run-tests.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/dev/run-tests.py b/dev/run-tests.py index c93423e36c92..ea515708124d 100755 --- a/dev/run-tests.py +++ b/dev/run-tests.py @@ -637,7 +637,7 @@ def main(): build_spark_assembly_sbt(hadoop_version, should_run_java_style_checks) # run the test suites - # run_scala_tests(build_tool, hadoop_version, test_modules, excluded_tags) + run_scala_tests(build_tool, hadoop_version, test_modules, excluded_tags) modules_with_python_tests = [m for m in test_modules if m.python_test_goals] if modules_with_python_tests: