-
Notifications
You must be signed in to change notification settings - Fork 241
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fixed some of the failing parquet_tests [databricks] #11429
Changes from 2 commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -32,7 +32,7 @@ import org.apache.spark.sql.catalyst.util.{truncatedString, InternalRowComparabl | |
import org.apache.spark.sql.connector.catalog.Table | ||
import org.apache.spark.sql.connector.read._ | ||
import org.apache.spark.sql.execution.datasources.rapids.DataSourceStrategyUtils | ||
import org.apache.spark.sql.execution.datasources.v2.{DataSourceRDD, StoragePartitionJoinParams} | ||
import org.apache.spark.sql.execution.datasources.v2._ | ||
import org.apache.spark.sql.internal.SQLConf | ||
|
||
case class GpuBatchScanExec( | ||
|
@@ -47,7 +47,7 @@ case class GpuBatchScanExec( | |
@transient override lazy val batch: Batch = if (scan == null) null else scan.toBatch | ||
// TODO: unify the equal/hashCode implementation for all data source v2 query plans. | ||
override def equals(other: Any): Boolean = other match { | ||
case other: GpuBatchScanExec => | ||
case other: BatchScanExec => | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Why is this being changed? This is a GpuBatchScanExec. We don't want to be equal to non-GPU versions do we? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Right. While debugging I wasn't sure what was causing a failure and looking at 330 shim I changed this and didn't change it back before submitting this PR. I am adding that change to this PR as well. |
||
this.batch != null && this.batch == other.batch && | ||
this.runtimeFilters == other.runtimeFilters && | ||
this.spjParams == other.spjParams | ||
|
@@ -137,6 +137,7 @@ case class GpuBatchScanExec( | |
override lazy val readerFactory: PartitionReaderFactory = batch.createReaderFactory() | ||
|
||
override lazy val inputRDD: RDD[InternalRow] = { | ||
scan.metrics = allMetrics | ||
val rdd = if (filteredPartitions.isEmpty && outputPartitioning == SinglePartition) { | ||
// return an empty RDD with 1 partition if dynamic filtering removed the only split | ||
sparkContext.parallelize(Array.empty[InternalRow], 1) | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
All of the non-legacy versions of these configs appear to have been added in 3.0.0. Is there a reason we are not just switching over to using them instead?