-
Notifications
You must be signed in to change notification settings - Fork 237
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support spark.sql.parquet.int96RebaseModeInWrite=LEGACY
[databricks]
#9658
Conversation
Signed-off-by: Nghia Truong <[email protected]>
# Conflicts: # sql-plugin/src/main/scala/com/nvidia/spark/RebaseHelper.scala # sql-plugin/src/main/scala/com/nvidia/spark/rapids/GpuParquetScan.scala
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
…IA#9617)" This reverts commit 401d0d8. Signed-off-by: Nghia Truong <[email protected]> # Conflicts: # sql-plugin/src/main/scala/com/nvidia/spark/rapids/GpuParquetScan.scala
# Conflicts: # sql-plugin/src/main/scala/com/nvidia/spark/RebaseHelper.scala # sql-plugin/src/main/scala/com/nvidia/spark/rapids/GpuParquetScan.scala
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
build |
build |
build |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I see the tests are updated and the code says that we now support rebase for in96 writes. But I don't see anywhere that the code was updated for it. I am assuming that the existing code just covered it and we are now enabling it after testing.
Signed-off-by: Nghia Truong <[email protected]>
Signed-off-by: Nghia Truong <[email protected]>
build |
Right, the existing code already handles the rebase computation. Now we just enable the corresponding code path and update tests. |
Signed-off-by: Nghia Truong <[email protected]>
build |
@@ -85,8 +85,6 @@ def do_write(spark, table_name): | |||
@pytest.mark.skipif(not is_hive_available(), reason="Hive is missing") | |||
@pytest.mark.parametrize("gens", [_basic_gens], ids=idfn) | |||
@pytest.mark.parametrize("storage_with_confs", [ | |||
("PARQUET", {"spark.sql.legacy.parquet.datetimeRebaseModeInWrite": "LEGACY", | |||
"spark.sql.legacy.parquet.int96RebaseModeInWrite": "LEGACY"}), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why drop these?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is a fallback test. We now have full support for LEGACY
in write thus we don't fallback anymore.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This adds support for
LEGACY
mode inspark.sql.parquet.int96RebaseModeInWrite
, which allows writing files containing ancient times before1582-10-15
with rebasing from Proleptic Gregorian calendar times to Julian calendar times.Closes: