Skip to content

Commit 159b3d5

Browse files
author
云峤
committed
update
1 parent 03ef434 commit 159b3d5

File tree

1 file changed

+13
-11
lines changed

1 file changed

+13
-11
lines changed

sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala

Lines changed: 13 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -20,31 +20,33 @@ package org.apache.spark.sql
2020
import java.io.CharArrayWriter
2121
import java.sql.DriverManager
2222

23+
24+
import scala.collection.JavaConversions._
25+
import scala.language.implicitConversions
26+
import scala.reflect.ClassTag
27+
import scala.reflect.runtime.universe.TypeTag
28+
import scala.util.control.NonFatal
29+
2330
import com.fasterxml.jackson.core.JsonFactory
31+
2432
import org.apache.commons.lang3.StringUtils
2533
import org.apache.spark.annotation.{DeveloperApi, Experimental}
2634
import org.apache.spark.api.java.JavaRDD
2735
import org.apache.spark.api.python.SerDeUtil
2836
import org.apache.spark.rdd.RDD
29-
import org.apache.spark.sql.catalyst.analysis.{ResolvedStar, UnresolvedAttribute, UnresolvedRelation}
37+
import org.apache.spark.storage.StorageLevel
38+
import org.apache.spark.sql.catalyst.{CatalystTypeConverters, ScalaReflection, SqlParser}
39+
import org.apache.spark.sql.catalyst.analysis.{UnresolvedAttribute, UnresolvedRelation, ResolvedStar}
3040
import org.apache.spark.sql.catalyst.expressions._
41+
import org.apache.spark.sql.catalyst.plans.{JoinType, Inner}
3142
import org.apache.spark.sql.catalyst.plans.logical._
32-
import org.apache.spark.sql.catalyst.plans.{Inner, JoinType}
33-
import org.apache.spark.sql.catalyst.{CatalystTypeConverters, ScalaReflection, SqlParser}
3443
import org.apache.spark.sql.execution.{EvaluatePython, ExplainCommand, LogicalRDD}
3544
import org.apache.spark.sql.jdbc.JDBCWriteDetails
3645
import org.apache.spark.sql.json.JsonRDD
37-
import org.apache.spark.sql.sources.{CreateTableUsingAsSelect, ResolvedDataSource}
3846
import org.apache.spark.sql.types._
39-
import org.apache.spark.storage.StorageLevel
47+
import org.apache.spark.sql.sources.{ResolvedDataSource, CreateTableUsingAsSelect}
4048
import org.apache.spark.util.Utils
4149

42-
import scala.collection.JavaConversions._
43-
import scala.language.implicitConversions
44-
import scala.reflect.ClassTag
45-
import scala.reflect.runtime.universe.TypeTag
46-
import scala.util.control.NonFatal
47-
4850

4951
private[sql] object DataFrame {
5052
def apply(sqlContext: SQLContext, logicalPlan: LogicalPlan): DataFrame = {

0 commit comments

Comments
 (0)