|
17 | 17 |
|
18 | 18 | package org.apache.spark.sql.hive |
19 | 19 |
|
20 | | -import org.apache.spark.sql.catalyst.expressions.Row |
21 | | - |
22 | 20 | import scala.collection.JavaConversions._ |
23 | 21 |
|
24 | 22 | import org.apache.spark.annotation.Experimental |
@@ -212,18 +210,22 @@ private[hive] trait HiveStrategies { |
212 | 210 |
|
213 | 211 | object HiveDDLStrategy extends Strategy { |
214 | 212 | def apply(plan: LogicalPlan): Seq[SparkPlan] = plan match { |
215 | | - case CreateTableUsing(tableIdentifier, userSpecifiedSchema, provider, false, opts, allowExisting) => |
| 213 | + case CreateTableUsing( |
| 214 | + tableIdentifier, userSpecifiedSchema, provider, false, opts, allowExisting) => |
216 | 215 | ExecutedCommand( |
217 | 216 | CreateMetastoreDataSource( |
218 | 217 | tableIdentifier, userSpecifiedSchema, provider, opts, allowExisting)) :: Nil |
219 | 218 |
|
220 | | - case CreateTableUsingAsSelect(tableIdentifier, provider, false, opts, allowExisting, query) => |
| 219 | + case CreateTableUsingAsSelect( |
| 220 | + tableIdentifier, provider, false, opts, allowExisting, query) => |
221 | 221 | val logicalPlan = hiveContext.parseSql(query) |
222 | 222 | val cmd = |
223 | | - CreateMetastoreDataSourceAsSelect(tableIdentifier, provider, opts, allowExisting, logicalPlan) |
| 223 | + CreateMetastoreDataSourceAsSelect( |
| 224 | + tableIdentifier, provider, opts, allowExisting, logicalPlan) |
224 | 225 | ExecutedCommand(cmd) :: Nil |
225 | 226 |
|
226 | | - case CreateTableUsingAsLogicalPlan(tableIdentifier, provider, false, opts, allowExisting, query) => |
| 227 | + case CreateTableUsingAsLogicalPlan( |
| 228 | + tableIdentifier, provider, false, opts, allowExisting, query) => |
227 | 229 | val cmd = |
228 | 230 | CreateMetastoreDataSourceAsSelect(tableIdentifier, provider, opts, allowExisting, query) |
229 | 231 | ExecutedCommand(cmd) :: Nil |
|
0 commit comments