From 445b57bc8dca3753be5dfeecc964be072933898a Mon Sep 17 00:00:00 2001 From: scwf Date: Wed, 31 Dec 2014 07:12:41 +0800 Subject: [PATCH] address comments --- .../main/scala/org/apache/spark/sql/sources/ddl.scala | 11 +++++++---- .../apache/spark/sql/sources/FilteredScanSuite.scala | 5 ++--- .../apache/spark/sql/sources/PrunedScanSuite.scala | 5 ++--- .../org/apache/spark/sql/sources/TableScanSuite.scala | 5 ++--- 4 files changed, 13 insertions(+), 13 deletions(-) diff --git a/sql/core/src/main/scala/org/apache/spark/sql/sources/ddl.scala b/sql/core/src/main/scala/org/apache/spark/sql/sources/ddl.scala index 0fb8597ded062..e5caa1c84c90b 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/sources/ddl.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/sources/ddl.scala @@ -209,12 +209,15 @@ private[sql] case class CreateTableUsing( sys.error(s"Failed to load class for data source: $provider") } } - val dataSource = - clazz.newInstance().asInstanceOf[org.apache.spark.sql.sources.SchemaRelationProvider] val relation = if(tableCols.isEmpty) { - dataSource.createRelation( - sqlContext, new CaseInsensitiveMap(options)) + val dataSource = + clazz.newInstance().asInstanceOf[org.apache.spark.sql.sources.RelationProvider] + + dataSource.createRelation(sqlContext, new CaseInsensitiveMap(options)) } else { + val dataSource = + clazz.newInstance().asInstanceOf[org.apache.spark.sql.sources.SchemaRelationProvider] + dataSource.createRelation( sqlContext, new CaseInsensitiveMap(options), Some(StructType(tableCols))) } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/FilteredScanSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/FilteredScanSuite.scala index 7ebcc1227e5cb..939b3c0c66de7 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/FilteredScanSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/FilteredScanSuite.scala @@ -21,11 +21,10 @@ import scala.language.existentials import org.apache.spark.sql._ -class FilteredScanSource extends SchemaRelationProvider { +class FilteredScanSource extends RelationProvider { override def createRelation( sqlContext: SQLContext, - parameters: Map[String, String], - schema: Option[StructType] = None): BaseRelation = { + parameters: Map[String, String]): BaseRelation = { SimpleFilteredScan(parameters("from").toInt, parameters("to").toInt)(sqlContext) } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/PrunedScanSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/PrunedScanSuite.scala index db6684ba22ab8..fee2e22611cdc 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/PrunedScanSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/PrunedScanSuite.scala @@ -19,11 +19,10 @@ package org.apache.spark.sql.sources import org.apache.spark.sql._ -class PrunedScanSource extends SchemaRelationProvider { +class PrunedScanSource extends RelationProvider { override def createRelation( sqlContext: SQLContext, - parameters: Map[String, String], - schema: Option[StructType] = None): BaseRelation = { + parameters: Map[String, String]): BaseRelation = { SimplePrunedScan(parameters("from").toInt, parameters("to").toInt)(sqlContext) } } diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/TableScanSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/TableScanSuite.scala index 63e2648178e22..3cd7b0115d567 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/TableScanSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/TableScanSuite.scala @@ -21,11 +21,10 @@ import org.apache.spark.sql._ class DefaultSource extends SimpleScanSource -class SimpleScanSource extends SchemaRelationProvider { +class SimpleScanSource extends RelationProvider { override def createRelation( sqlContext: SQLContext, - parameters: Map[String, String], - schema: Option[StructType] = None): BaseRelation = { + parameters: Map[String, String]): BaseRelation = { SimpleScan(parameters("from").toInt, parameters("TO").toInt)(sqlContext) } }