diff --git a/docs/sql-data-sources-parquet.md b/docs/sql-data-sources-parquet.md index 119eae56ebf7..de278cf9f924 100644 --- a/docs/sql-data-sources-parquet.md +++ b/docs/sql-data-sources-parquet.md @@ -255,8 +255,11 @@ REFRESH TABLE my_table; ## Data Source Option Data source options of Parquet can be set via: -* the `.option`/`.options` methods of `DataFrameReader` or `DataFrameWriter` -* the `.option`/`.options` methods of `DataStreamReader` or `DataStreamWriter` +* the `.option`/`.options` methods of + * `DataFrameReader` + * `DataFrameWriter` + * `DataStreamReader` + * `DataStreamWriter`
| Property Name | Default | Meaning | Scope | read | +
|---|---|---|---|
mergeSchema |
+ The SQL config spark.sql.parquet.mergeSchema which is false by default. |
+ Sets whether we should merge schemas collected from all Parquet part-files. This will override spark.sql.parquet.mergeSchema. |
+ read | +
compression |
+ None | +Compression codec to use when saving to file. This can be one of the known case-insensitive shorten names (none, uncompressed, snappy, gzip, lzo, brotli, lz4, and zstd). This will override spark.sql.parquet.compression.codec. If None is set, it uses the value specified in spark.sql.parquet.compression.codec. |
+ write | +
org.apache.hadoop.fs.GlobFilter.
- * It does not change the behavior of partition discovery.org.apache.hadoop.fs.GlobFilter.
- * It does not change the behavior of partition discovery.