File tree Expand file tree Collapse file tree 2 files changed +2
-7
lines changed Expand file tree Collapse file tree 2 files changed +2
-7
lines changed Original file line number Diff line number Diff line change 2121 <parent >
2222 <groupId >org.apache.spark</groupId >
2323 <artifactId >spark-parent</artifactId >
24- <<<<<<< HEAD
2524 <version >1.2.0-SNAPSHOT</version >
26- =======
27- <version >1.1.0-SNAPSHOT</version >
28- >>>>>>> modified the code base on comment in https://github.com/tdas/spark/pull/10
2925 <relativePath >../pom.xml</relativePath >
3026 </parent >
3127
Original file line number Diff line number Diff line change 2222from pyspark .serializers import NoOpSerializer ,\
2323 BatchedSerializer , CloudPickleSerializer , pack_long ,\
2424 CompressedSerializer
25- from pyspark .rdd import _JavaStackTrace
2625from pyspark .storagelevel import StorageLevel
2726from pyspark .resultiterable import ResultIterable
2827from pyspark .streaming .util import rddToFileName , RDDFunction
29-
28+ from pyspark . traceback_utils import SCCallSiteSync
3029
3130from py4j .java_collections import ListConverter , MapConverter
3231
@@ -187,7 +186,7 @@ def add_shuffle_key(split, iterator):
187186 yield outputSerializer .dumps (items )
188187 keyed = PipelinedDStream (self , add_shuffle_key )
189188 keyed ._bypass_serializer = True
190- with _JavaStackTrace (self .ctx ) as st :
189+ with SCCallSiteSync (self .context ) as css :
191190 partitioner = self .ctx ._jvm .PythonPartitioner (numPartitions ,
192191 id (partitionFunc ))
193192 jdstream = self .ctx ._jvm .PythonPairwiseDStream (keyed ._jdstream .dstream (),
You can’t perform that action at this time.
0 commit comments