-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-5549] Define TaskContext interface in Scala. #4324
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -15,112 +15,116 @@ | |
| * limitations under the License. | ||
| */ | ||
|
|
||
| package org.apache.spark; | ||
| package org.apache.spark | ||
|
|
||
| import java.io.Serializable; | ||
| import java.io.Serializable | ||
|
|
||
| import scala.Function0; | ||
| import scala.Function1; | ||
| import scala.Unit; | ||
| import org.apache.spark.annotation.DeveloperApi | ||
| import org.apache.spark.executor.TaskMetrics | ||
| import org.apache.spark.util.TaskCompletionListener | ||
|
|
||
| import org.apache.spark.annotation.DeveloperApi; | ||
| import org.apache.spark.executor.TaskMetrics; | ||
| import org.apache.spark.util.TaskCompletionListener; | ||
|
|
||
| /** | ||
| * Contextual information about a task which can be read or mutated during | ||
| * execution. To access the TaskContext for a running task use | ||
| * TaskContext.get(). | ||
| */ | ||
| public abstract class TaskContext implements Serializable { | ||
| object TaskContext { | ||
| /** | ||
| * Return the currently active TaskContext. This can be called inside of | ||
| * user functions to access contextual information about running tasks. | ||
| */ | ||
| public static TaskContext get() { | ||
| return taskContext.get(); | ||
| } | ||
| def get(): TaskContext = taskContext.get | ||
|
|
||
| private val taskContext: ThreadLocal[TaskContext] = new ThreadLocal[TaskContext] | ||
|
|
||
| private[spark] def setTaskContext(tc: TaskContext): Unit = taskContext.set(tc) | ||
|
|
||
| private[spark] def unset(): Unit = taskContext.remove() | ||
| } | ||
|
|
||
| private static ThreadLocal<TaskContext> taskContext = | ||
| new ThreadLocal<TaskContext>(); | ||
|
|
||
| static void setTaskContext(TaskContext tc) { | ||
| taskContext.set(tc); | ||
| } | ||
| /** | ||
| * Contextual information about a task which can be read or mutated during | ||
| * execution. To access the TaskContext for a running task, use: | ||
| * {{{ | ||
| * org.apache.spark.TaskContext.get() | ||
| * }}} | ||
| */ | ||
| abstract class TaskContext extends Serializable { | ||
| // Note: TaskContext must NOT define a get method. Otherwise it will prevent the Scala compiler | ||
| // from generating a static get method (based on the companion object's get method). | ||
|
|
||
| static void unset() { | ||
| taskContext.remove(); | ||
| } | ||
| // Note: getters in this class are defined with parentheses to maintain backward compatibility. | ||
|
|
||
| /** | ||
| * Whether the task has completed. | ||
| * Returns true if the task has completed. | ||
| */ | ||
| public abstract boolean isCompleted(); | ||
| def isCompleted(): Boolean | ||
|
|
||
| /** | ||
| * Whether the task has been killed. | ||
| * Returns true if the task has been killed. | ||
| */ | ||
| public abstract boolean isInterrupted(); | ||
| def isInterrupted(): Boolean | ||
|
|
||
| /** @deprecated use {@link #isRunningLocally()} */ | ||
| @Deprecated | ||
| public abstract boolean runningLocally(); | ||
| /** @deprecated use { @link #isRunningLocally()}*/ | ||
| @deprecated("1.2.0", "use isRunningLocally") | ||
| def runningLocally(): Boolean | ||
|
|
||
| public abstract boolean isRunningLocally(); | ||
| /** | ||
| * Returns true if the task is running locally in the driver program. | ||
| * @return | ||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Can probably drop this |
||
| */ | ||
| def isRunningLocally(): Boolean | ||
|
|
||
| /** | ||
| * Add a (Java friendly) listener to be executed on task completion. | ||
| * Adds a (Java friendly) listener to be executed on task completion. | ||
| * This will be called in all situation - success, failure, or cancellation. | ||
| * An example use is for HadoopRDD to register a callback to close the input stream. | ||
| */ | ||
| public abstract TaskContext addTaskCompletionListener(TaskCompletionListener listener); | ||
| def addTaskCompletionListener(listener: TaskCompletionListener): TaskContext | ||
|
|
||
| /** | ||
| * Add a listener in the form of a Scala closure to be executed on task completion. | ||
| * Adds a listener in the form of a Scala closure to be executed on task completion. | ||
| * This will be called in all situations - success, failure, or cancellation. | ||
| * An example use is for HadoopRDD to register a callback to close the input stream. | ||
| */ | ||
| public abstract TaskContext addTaskCompletionListener(final Function1<TaskContext, Unit> f); | ||
| def addTaskCompletionListener(f: (TaskContext) => Unit): TaskContext | ||
|
|
||
| /** | ||
| * Add a callback function to be executed on task completion. An example use | ||
| * Adds a callback function to be executed on task completion. An example use | ||
| * is for HadoopRDD to register a callback to close the input stream. | ||
| * Will be called in any situation - success, failure, or cancellation. | ||
| * | ||
| * @deprecated use {@link #addTaskCompletionListener(scala.Function1)} | ||
| * @deprecated use { @link #addTaskCompletionListener(scala.Function1)} | ||
| * | ||
| * @param f Callback function. | ||
| */ | ||
| @Deprecated | ||
| public abstract void addOnCompleteCallback(final Function0<Unit> f); | ||
| @deprecated("1.2.0", "use addTaskCompletionListener") | ||
| def addOnCompleteCallback(f: () => Unit) | ||
|
|
||
| /** | ||
| * The ID of the stage that this task belong to. | ||
| */ | ||
| public abstract int stageId(); | ||
| def stageId(): Int | ||
|
|
||
| /** | ||
| * The ID of the RDD partition that is computed by this task. | ||
| */ | ||
| public abstract int partitionId(); | ||
| def partitionId(): Int | ||
|
|
||
| /** | ||
| * How many times this task has been attempted. The first task attempt will be assigned | ||
| * attemptNumber = 0, and subsequent attempts will have increasing attempt numbers. | ||
| */ | ||
| public abstract int attemptNumber(); | ||
| def attemptNumber(): Int | ||
|
|
||
| /** @deprecated use {@link #taskAttemptId()}; it was renamed to avoid ambiguity. */ | ||
| @Deprecated | ||
| public abstract long attemptId(); | ||
| /** @deprecated use { @link #taskAttemptId()}; it was renamed to avoid ambiguity. */ | ||
| @deprecated("1.3.0", "use attemptNumber") | ||
| def attemptId(): Long | ||
|
|
||
| /** | ||
| * An ID that is unique to this task attempt (within the same SparkContext, no two task attempts | ||
| * will share the same attempt ID). This is roughly equivalent to Hadoop's TaskAttemptID. | ||
| */ | ||
| public abstract long taskAttemptId(); | ||
| def taskAttemptId(): Long | ||
|
|
||
| /** ::DeveloperApi:: */ | ||
| @DeveloperApi | ||
| public abstract TaskMetrics taskMetrics(); | ||
| def taskMetrics(): TaskMetrics | ||
| } | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -15,9 +15,10 @@ | |
| * limitations under the License. | ||
| */ | ||
|
|
||
| package org.apache.spark.util; | ||
| package test.org.apache.spark; | ||
|
Contributor
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. note that I changed the package to make sure we test package visibility correctly also. We should do this for other Java API tests, but we can do those later. |
||
|
|
||
| import org.apache.spark.TaskContext; | ||
| import org.apache.spark.util.TaskCompletionListener; | ||
|
|
||
|
|
||
| /** | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,41 @@ | ||
| /* | ||
| * Licensed to the Apache Software Foundation (ASF) under one or more | ||
| * contributor license agreements. See the NOTICE file distributed with | ||
| * this work for additional information regarding copyright ownership. | ||
| * The ASF licenses this file to You under the Apache License, Version 2.0 | ||
| * (the "License"); you may not use this file except in compliance with | ||
| * the License. You may obtain a copy of the License at | ||
| * | ||
| * http://www.apache.org/licenses/LICENSE-2.0 | ||
| * | ||
| * Unless required by applicable law or agreed to in writing, software | ||
| * distributed under the License is distributed on an "AS IS" BASIS, | ||
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||
| * See the License for the specific language governing permissions and | ||
| * limitations under the License. | ||
| */ | ||
|
|
||
| package test.org.apache.spark; | ||
|
|
||
| import org.apache.spark.TaskContext; | ||
|
|
||
| /** | ||
| * Something to make sure that TaskContext can be used in Java. | ||
| */ | ||
| public class JavaTaskContextCompileCheck { | ||
|
|
||
| public static void test() { | ||
| TaskContext tc = TaskContext.get(); | ||
|
|
||
| tc.isCompleted(); | ||
| tc.isInterrupted(); | ||
| tc.isRunningLocally(); | ||
|
|
||
| tc.addTaskCompletionListener(new JavaTaskCompletionListenerImpl()); | ||
|
|
||
| tc.attemptNumber(); | ||
| tc.partitionId(); | ||
| tc.stageId(); | ||
| tc.taskAttemptId(); | ||
| } | ||
| } |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Scalastyle didn't like the whitespace here.