Skip to content

Commit

Permalink
Add GPU metrics to GpuFileSourceScanExec (#547)
Browse files Browse the repository at this point in the history
* Add GPU metrics to GpuFileSourceScanExec

Signed-off-by: Jason Lowe <[email protected]>

* Extract GpuFileSourceScanExec from shims

Signed-off-by: Jason Lowe <[email protected]>

* Pass metrics via GPU file format rather than custom options map

Signed-off-by: Jason Lowe <[email protected]>

* Update code checking for DataSourceScanExec

Signed-off-by: Jason Lowe <[email protected]>

* Fix scaladoc warning and unused imports

Signed-off-by: Jason Lowe <[email protected]>

* Fix copyright

Signed-off-by: Jason Lowe <[email protected]>
  • Loading branch information
jlowe authored Aug 14, 2020
1 parent b86fd32 commit 20afca1
Show file tree
Hide file tree
Showing 16 changed files with 728 additions and 628 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ import org.apache.spark.sql.execution.datasources.HadoopFsRelation
import org.apache.spark.sql.execution.exchange.{BroadcastExchangeExec, ShuffleExchangeExec}
import org.apache.spark.sql.execution.joins.{BroadcastHashJoinExec, BroadcastNestedLoopJoinExec, HashJoin, SortMergeJoinExec}
import org.apache.spark.sql.execution.joins.ShuffledHashJoinExec
import org.apache.spark.sql.rapids.{GpuTimeSub, ShuffleManagerShimBase}
import org.apache.spark.sql.rapids.{GpuFileSourceScanExec, GpuTimeSub, ShuffleManagerShimBase}
import org.apache.spark.sql.rapids.execution.{GpuBroadcastExchangeExecBase, GpuBroadcastNestedLoopJoinExecBase, GpuShuffleExchangeExecBase}
import org.apache.spark.sql.rapids.shims.spark300._
import org.apache.spark.sql.types._
Expand Down Expand Up @@ -148,6 +148,7 @@ class Spark300Shims extends SparkShims {
wrapped.requiredSchema,
wrapped.partitionFilters,
wrapped.optionalBucketSet,
None,
wrapped.dataFilters,
wrapped.tableIdentifier)
}
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@ import org.apache.spark.sql.execution._
import org.apache.spark.sql.execution.datasources.HadoopFsRelation
import org.apache.spark.sql.execution.joins.{BroadcastHashJoinExec, BroadcastNestedLoopJoinExec, HashJoin, SortMergeJoinExec}
import org.apache.spark.sql.execution.joins.ShuffledHashJoinExec
import org.apache.spark.sql.rapids.GpuTimeSub
import org.apache.spark.sql.rapids.{GpuFileSourceScanExec, GpuTimeSub}
import org.apache.spark.sql.rapids.execution.GpuBroadcastNestedLoopJoinExecBase
import org.apache.spark.sql.rapids.shims.spark300db._
import org.apache.spark.sql.types._
Expand Down Expand Up @@ -94,6 +94,8 @@ class Spark300dbShims extends Spark300Shims {
wrapped.requiredSchema,
wrapped.partitionFilters,
wrapped.optionalBucketSet,
// TODO: Does Databricks have coalesced bucketing implemented?
None,
wrapped.dataFilters,
wrapped.tableIdentifier)
}
Expand Down

This file was deleted.

Loading

0 comments on commit 20afca1

Please sign in to comment.