Skip to content

Commit 92e017f

Browse files
scwfJoshRosen
authored andcommitted
[SPARK-3899][Doc]fix wrong links in streaming doc
There are three [Custom Receiver Guide] links in streaming doc, the first is wrong. Author: w00228970 <[email protected]> Author: wangfei <[email protected]> Closes apache#2749 from scwf/streaming-doc and squashes the following commits: 0cd76b7 [wangfei] update link tojump to the Akka-specific section 45b0646 [w00228970] wrong link in streaming doc
1 parent d8b8c21 commit 92e017f

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/streaming-programming-guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -494,7 +494,7 @@ methods for creating DStreams from files and Akka actors as input sources.
494494

495495
For simple text files, there is an easier method `streamingContext.textFileStream(dataDirectory)`. And file streams do not require running a receiver, hence does not require allocating cores.
496496

497-
- **Streams based on Custom Actors:** DStreams can be created with data streams received through Akka actors by using `streamingContext.actorStream(actorProps, actor-name)`. See the [Custom Receiver Guide](#implementing-and-using-a-custom-actor-based-receiver) for more details.
497+
- **Streams based on Custom Actors:** DStreams can be created with data streams received through Akka actors by using `streamingContext.actorStream(actorProps, actor-name)`. See the [Custom Receiver Guide](streaming-custom-receivers.html#implementing-and-using-a-custom-actor-based-receiver) for more details.
498498

499499
- **Queue of RDDs as a Stream:** For testing a Spark Streaming application with test data, one can also create a DStream based on a queue of RDDs, using `streamingContext.queueStream(queueOfRDDs)`. Each RDD pushed into the queue will be treated as a batch of data in the DStream, and processed like a stream.
500500

0 commit comments

Comments
 (0)