diff --git a/python/pyspark/sql/readwriter.py b/python/pyspark/sql/readwriter.py index d555bad75a3ea..1353a0f6e9107 100644 --- a/python/pyspark/sql/readwriter.py +++ b/python/pyspark/sql/readwriter.py @@ -509,8 +509,6 @@ def func(iterator): def orc(self, path): """Loads ORC files, returning the result as a :class:`DataFrame`. - .. note:: Currently ORC support is only available together with Hive support. - >>> df = spark.read.orc('python/test_support/sql/orc_partitioned') >>> df.dtypes [('a', 'bigint'), ('b', 'int'), ('c', 'int')] @@ -950,8 +948,6 @@ def csv(self, path, mode=None, compression=None, sep=None, quote=None, escape=No def orc(self, path, mode=None, partitionBy=None, compression=None): """Saves the content of the :class:`DataFrame` in ORC format at the specified path. - .. note:: Currently ORC support is only available together with Hive support. - :param path: the path in any Hadoop supported file system :param mode: specifies the behavior of the save operation when data already exists. diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala index d514b9495c5d8..3cd48baebf2e3 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameReader.scala @@ -679,7 +679,6 @@ class DataFrameReader private[sql](sparkSession: SparkSession) extends Logging { * * @param path input path * @since 1.5.0 - * @note Currently, this method can only be used after enabling Hive support. */ def orc(path: String): DataFrame = { // This method ensures that calls that explicit need single argument works, see SPARK-16009 @@ -691,7 +690,6 @@ class DataFrameReader private[sql](sparkSession: SparkSession) extends Logging { * * @param paths input paths * @since 2.0.0 - * @note Currently, this method can only be used after enabling Hive support. */ @scala.annotation.varargs def orc(paths: String*): DataFrame = format("orc").load(paths: _*) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala index a06b75a860679..9371936047984 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala @@ -615,7 +615,6 @@ final class DataFrameWriter[T] private[sql](ds: Dataset[T]) { * * * @since 1.5.0 - * @note Currently, this method can only be used after enabling Hive support */ def orc(path: String): Unit = { format("orc").save(path)