diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcUtils.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcUtils.scala index e10253989788b..072e670081d1b 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcUtils.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/orc/OrcUtils.scala @@ -207,10 +207,16 @@ object OrcUtils extends Logging { } /** - * @return Returns the result schema string based on the canPruneCols flag. - * resultSchemaString will be created using resultsSchema in case of - * canPruneCols is true and for canPruneCols as false value - * resultSchemaString will be created using the actual dataSchema. + * Returns the result schema to read from ORC file. In addition, It sets + * the schema string to 'orc.mapred.input.schema' so ORC reader can use later. + * + * @param canPruneCols Flag to decide whether pruned cols schema is send to resultSchema + * or to send the entire dataSchema to resultSchema. + * @param dataSchema Schema of the orc files. + * @param resultSchema Result data schema created after pruning cols. + * @param partitionSchema Schema of partitions. + * @param conf Hadoop Configuration. + * @return Returns the result schema as string. */ def orcResultSchemaString( canPruneCols: Boolean,