1717
1818package org .apache .spark .sql .hive
1919
20- import java .io .IOException
21- import java .util .{List => JList }
22-
2320import com .google .common .base .Objects
2421import com .google .common .cache .{CacheBuilder , CacheLoader , LoadingCache }
2522
2623import org .apache .hadoop .fs .Path
27- import org .apache .hadoop .hive .metastore .api .{ FieldSchema , Partition => TPartition , Table => TTable }
28- import org .apache .hadoop .hive .metastore .{ TableType , Warehouse }
24+ import org .apache .hadoop .hive .metastore .api .FieldSchema
25+ import org .apache .hadoop .hive .metastore .Warehouse
2926import org .apache .hadoop .hive .ql .metadata ._
30- import org .apache .hadoop .hive .ql .plan .CreateTableDesc
31- import org .apache .hadoop .hive .serde .serdeConstants
32- import org .apache .hadoop .hive .serde2 .`lazy` .LazySimpleSerDe
33- import org .apache .hadoop .hive .serde2 .{Deserializer , SerDeException }
34- import org .apache .hadoop .util .ReflectionUtils
27+ import org .apache .hadoop .hive .serde2 .Deserializer
3528
3629import org .apache .spark .Logging
37- import org .apache .spark .sql .hive .client .IsolatedClientLoader
3830import org .apache .spark .sql .{SaveMode , AnalysisException , SQLContext }
39- import org .apache .spark .sql .catalyst .analysis .{MultiInstanceRelation , NoSuchTableException , Catalog , OverrideCatalog }
31+ import org .apache .spark .sql .catalyst .analysis .{MultiInstanceRelation , Catalog , OverrideCatalog }
4032import org .apache .spark .sql .catalyst .expressions ._
4133import org .apache .spark .sql .catalyst .planning .PhysicalOperation
4234import org .apache .spark .sql .catalyst .plans .logical
4335import org .apache .spark .sql .catalyst .plans .logical ._
4436import org .apache .spark .sql .catalyst .rules ._
4537import org .apache .spark .sql .hive .client ._
4638import org .apache .spark .sql .parquet .{ParquetRelation2 , Partition => ParquetPartition , PartitionSpec }
47- import org .apache .spark .sql .sources .{CreateTableUsingAsSelect , DDLParser , LogicalRelation , ResolvedDataSource }
39+ import org .apache .spark .sql .sources .{CreateTableUsingAsSelect , LogicalRelation , ResolvedDataSource }
4840import org .apache .spark .sql .types ._
49- import org .apache .spark .sql .catalyst .CatalystConf
5041import org .apache .spark .util .Utils
5142
5243/* Implicit conversions */
@@ -55,8 +46,6 @@ import scala.collection.JavaConversions._
5546private [hive] class HiveMetastoreCatalog (val client : ClientInterface , hive : HiveContext )
5647 extends Catalog with Logging {
5748
58- import org .apache .spark .sql .hive .HiveMetastoreTypes ._
59-
6049 /** Usages should lock on `this`. */
6150 protected [hive] lazy val hiveWarehouse = new Warehouse (hive.hiveconf)
6251
0 commit comments