diff --git a/core/src/main/scala/org/apache/spark/shuffle/sort/SortShuffleManager.scala b/core/src/main/scala/org/apache/spark/shuffle/sort/SortShuffleManager.scala index aefcb59b8bb87..353b82f9f6eef 100644 --- a/core/src/main/scala/org/apache/spark/shuffle/sort/SortShuffleManager.scala +++ b/core/src/main/scala/org/apache/spark/shuffle/sort/SortShuffleManager.scala @@ -154,7 +154,7 @@ private[spark] class SortShuffleManager(conf: SparkConf) extends ShuffleManager metrics: ShuffleWriteMetricsReporter): ShuffleWriter[K, V] = { val mapTaskIds = taskIdMapsForShuffle.computeIfAbsent( handle.shuffleId, _ => new OpenHashSet[Long](16)) - mapTaskIds.synchronized { mapTaskIds.add(context.taskAttemptId()) } + mapTaskIds.synchronized { mapTaskIds.add(mapId) } val env = SparkEnv.get handle match { case unsafeShuffleHandle: SerializedShuffleHandle[K @unchecked, V @unchecked] => diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/StringExpressionsSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/StringExpressionsSuite.scala index 967ccc42c632d..6468bf53ddb7d 100644 --- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/StringExpressionsSuite.scala +++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/StringExpressionsSuite.scala @@ -28,7 +28,7 @@ class StringExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper { test("concat") { def testConcat(inputs: String*): Unit = { val expected = if (inputs.contains(null)) null else inputs.mkString - checkEvaluation(Concat(inputs.map(Literal.create(_, StringType))), expected, EmptyRow) + checkEvaluation(Concat(inputs.map(Literal.create(_, StringType))), expected) } testConcat() @@ -50,7 +50,7 @@ class StringExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper { test("SPARK-22498: Concat should not generate codes beyond 64KB") { val N = 5000 val strs = (1 to N).map(x => s"s$x") - checkEvaluation(Concat(strs.map(Literal.create(_, StringType))), strs.mkString, EmptyRow) + checkEvaluation(Concat(strs.map(Literal.create(_, StringType))), strs.mkString) } test("SPARK-22771 Check Concat.checkInputDataTypes results") { @@ -73,7 +73,7 @@ class StringExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper { case s: String => Literal.create(s, StringType) } val sepExpr = Literal.create(sep, StringType) - checkEvaluation(ConcatWs(sepExpr +: inputExprs), expected, EmptyRow) + checkEvaluation(ConcatWs(sepExpr +: inputExprs), expected) } // scalastyle:off @@ -99,12 +99,12 @@ class StringExpressionsSuite extends SparkFunSuite with ExpressionEvalHelper { val sepExpr = Literal.create("#", StringType) val strings1 = (1 to N).map(x => s"s$x") val inputsExpr1 = strings1.map(Literal.create(_, StringType)) - checkEvaluation(ConcatWs(sepExpr +: inputsExpr1), strings1.mkString("#"), EmptyRow) + checkEvaluation(ConcatWs(sepExpr +: inputsExpr1), strings1.mkString("#")) val strings2 = (1 to N).map(x => Seq(s"s$x")) val inputsExpr2 = strings2.map(Literal.create(_, ArrayType(StringType))) checkEvaluation( - ConcatWs(sepExpr +: inputsExpr2), strings2.map(s => s(0)).mkString("#"), EmptyRow) + ConcatWs(sepExpr +: inputsExpr2), strings2.map(s => s(0)).mkString("#")) } test("elt") {