diff --git a/python/run-tests-with-coverage b/python/run-tests-with-coverage index d1c2dacbf9d8..aa23e16e8e43 100755 --- a/python/run-tests-with-coverage +++ b/python/run-tests-with-coverage @@ -44,9 +44,6 @@ export PYTHONPATH="$FWDIR:$PYTHONPATH" # Also, our sitecustomize.py and coverage_daemon.py are included in the path. export PYTHONPATH="$COVERAGE_DIR:$PYTHONPATH" -# We use 'spark.python.daemon.module' configuration to insert the coverage supported workers. -export SPARK_CONF_DIR="$COVERAGE_DIR/conf" - # This environment variable enables the coverage. export COVERAGE_PROCESS_START="$FWDIR/.coveragerc" diff --git a/python/test_coverage/conf/spark-defaults.conf b/python/test_coverage/conf/spark-defaults.conf deleted file mode 100644 index bf44ea6e7cfe..000000000000 --- a/python/test_coverage/conf/spark-defaults.conf +++ /dev/null @@ -1,21 +0,0 @@ -# -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# - -# This is used to generate PySpark coverage results. Seems there's no way to -# add a configuration when SPARK_TESTING environment variable is set because -# we will directly execute modules by python -m. -spark.python.daemon.module coverage_daemon diff --git a/python/test_coverage/coverage_daemon.py b/python/test_coverage/coverage_daemon.py deleted file mode 100644 index 4372135d6fc3..000000000000 --- a/python/test_coverage/coverage_daemon.py +++ /dev/null @@ -1,48 +0,0 @@ -# -# Licensed to the Apache Software Foundation (ASF) under one or more -# contributor license agreements. See the NOTICE file distributed with -# this work for additional information regarding copyright ownership. -# The ASF licenses this file to You under the Apache License, Version 2.0 -# (the "License"); you may not use this file except in compliance with -# the License. You may obtain a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, -# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -# See the License for the specific language governing permissions and -# limitations under the License. -# - -import os -import imp -import platform - - -# This is a hack to always refer the main code rather than built zip. -main_code_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__))) -daemon = imp.load_source("daemon", "%s/pyspark/daemon.py" % main_code_dir) - -if "COVERAGE_PROCESS_START" in os.environ: - # PyPy with coverage makes the tests flaky, and CPython is enough for coverage report. - if "pypy" not in platform.python_implementation().lower(): - worker = imp.load_source("worker", "%s/pyspark/worker.py" % main_code_dir) - - def _cov_wrapped(*args, **kwargs): - import coverage - cov = coverage.coverage( - config_file=os.environ["COVERAGE_PROCESS_START"]) - cov.start() - try: - worker.main(*args, **kwargs) - finally: - cov.stop() - cov.save() - daemon.worker_main = _cov_wrapped -else: - raise RuntimeError("COVERAGE_PROCESS_START environment variable is not set, exiting.") - - -if __name__ == '__main__': - daemon.manager()