diff --git a/docs/quick-start.md b/docs/quick-start.md index 7cb3f75132206..a85e5b28a6996 100644 --- a/docs/quick-start.md +++ b/docs/quick-start.md @@ -297,12 +297,13 @@ We'll create a very simple Spark application, `SimpleApp.java`: {% highlight java %} /* SimpleApp.java */ import org.apache.spark.sql.SparkSession; +import org.apache.spark.sql.Dataset; public class SimpleApp { public static void main(String[] args) { String logFile = "YOUR_SPARK_HOME/README.md"; // Should be some file on your system SparkSession spark = SparkSession.builder().appName("Simple Application").getOrCreate(); - Dataset logData = spark.read.textFile(logFile).cache(); + Dataset logData = spark.read().textFile(logFile).cache(); long numAs = logData.filter(s -> s.contains("a")).count(); long numBs = logData.filter(s -> s.contains("b")).count(); diff --git a/docs/structured-streaming-programming-guide.md b/docs/structured-streaming-programming-guide.md index 8f64faadc32dc..8367f5a08c755 100644 --- a/docs/structured-streaming-programming-guide.md +++ b/docs/structured-streaming-programming-guide.md @@ -1041,8 +1041,8 @@ streamingDf.join(staticDf, "type", "right_join") // right outer join with a sta
{% highlight java %} -Dataset staticDf = spark.read. ...; -Dataset streamingDf = spark.readStream. ...; +Dataset staticDf = spark.read(). ...; +Dataset streamingDf = spark.readStream(). ...; streamingDf.join(staticDf, "type"); // inner equi-join with a static DF streamingDf.join(staticDf, "type", "right_join"); // right outer join with a static DF {% endhighlight %} @@ -1087,7 +1087,7 @@ streamingDf
{% highlight java %} -Dataset streamingDf = spark.readStream. ...; // columns: guid, eventTime, ... +Dataset streamingDf = spark.readStream(). ...; // columns: guid, eventTime, ... // Without watermark using guid column streamingDf.dropDuplicates("guid");