Skip to content

Commit 08913ce

Browse files
Seigneurin, Alexis (CONT)srowen
authored andcommitted
fixed a typo
idempotant -> idempotent Author: Seigneurin, Alexis (CONT) <[email protected]> Closes #14833 from aseigneurin/fix-typo.
1 parent 1a48c00 commit 08913ce

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/structured-streaming-programming-guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -406,7 +406,7 @@ Furthermore, this model naturally handles data that has arrived later than expec
406406

407407
## Fault Tolerance Semantics
408408
Delivering end-to-end exactly-once semantics was one of key goals behind the design of Structured Streaming. To achieve that, we have designed the Structured Streaming sources, the sinks and the execution engine to reliably track the exact progress of the processing so that it can handle any kind of failure by restarting and/or reprocessing. Every streaming source is assumed to have offsets (similar to Kafka offsets, or Kinesis sequence numbers)
409-
to track the read position in the stream. The engine uses checkpointing and write ahead logs to record the offset range of the data being processed in each trigger. The streaming sinks are designed to be idempotent for handling reprocessing. Together, using replayable sources and idempotant sinks, Structured Streaming can ensure **end-to-end exactly-once semantics** under any failure.
409+
to track the read position in the stream. The engine uses checkpointing and write ahead logs to record the offset range of the data being processed in each trigger. The streaming sinks are designed to be idempotent for handling reprocessing. Together, using replayable sources and idempotent sinks, Structured Streaming can ensure **end-to-end exactly-once semantics** under any failure.
410410

411411
# API using Datasets and DataFrames
412412
Since Spark 2.0, DataFrames and Datasets can represent static, bounded data, as well as streaming, unbounded data. Similar to static Datasets/DataFrames, you can use the common entry point `SparkSession` ([Scala](api/scala/index.html#org.apache.spark.sql.SparkSession)/

0 commit comments

Comments
 (0)