-
Notifications
You must be signed in to change notification settings - Fork 237
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] test_dpp_reuse_broadcast_exchange failed #10147
Comments
From CI, this test failure occurred in Spark 3.5.0 |
I was able to replicate both failures on Spark 3.2.4, 3.3.3, 3.4.0, and 3.5.0 (all versions of Spark that support AQE + DPP) |
Basically by the plan output here, it looks like this is an AQE optimization that is turning the entire plan into a
|
I guess it determined via the join that this would return empty |
So basically after some debugging, I think one the subqueries returned an empty result, so that was short-circuited by AQE to return a |
I'm not sure that's the best fix. The point of this test is to check handling of DPP, and the problem here is that the datagen happened to produce inputs that failed to produce a plan requiring DPP. IMHO a better fix is to update the input data generation to ensure there isn't a degenerate join. If we want to test handling of degenerate joins as well, that should be a separate test that explicitly sets up inputs to produce a degenerate join. |
Makes sense. Will investigate what is producing the empty join |
|
Test is now failing again:
|
Saw this fail again on Dataproc nightly run.
|
Another failure
|
Considering this is actually a test issue (the test not being able to avoid an empty LocalTableScan) and not an issue with the plugin, lowering the priority |
From a recent nightly test run:
The text was updated successfully, but these errors were encountered: