From cbf44998e4a9071840e1010f2d4e526e7126383e Mon Sep 17 00:00:00 2001 From: Nghia Truong Date: Wed, 13 Nov 2024 21:02:10 -0800 Subject: [PATCH] Remove empty lines --- .../org/apache/spark/sql/rapids/GpuJsonReadCommon.scala | 5 ----- 1 file changed, 5 deletions(-) diff --git a/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/GpuJsonReadCommon.scala b/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/GpuJsonReadCommon.scala index 5a5081a31ae7..c02993e22864 100644 --- a/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/GpuJsonReadCommon.scala +++ b/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/GpuJsonReadCommon.scala @@ -130,13 +130,10 @@ object GpuJsonReadCommon { options: JSONOptions): ColumnVector = { ColumnCastUtil.deepTransform(inputCv, Some(topLevelType), Some(nestedColumnViewMismatchTransform)) { - case (cv, Some(DateType)) if cv.getType == DType.STRING => convertStringToDate(cv, options) - case (cv, Some(TimestampType)) if cv.getType == DType.STRING => convertStringToTimestamp(cv, options) - case (cv, Some(dt)) if cv.getType == DType.STRING => // There is an issue with the Schema implementation such that the schema's top level // is never used when passing down data schema from Java to C++. @@ -180,10 +177,8 @@ object GpuJsonReadCommon { withResource(new NvtxRange("convertDateTimeType", NvtxColor.RED)) { _ => ColumnCastUtil.deepTransform(inputCv, Some(topLevelType), Some(nestedColumnViewMismatchTransform)) { - case (cv, Some(DateType)) if cv.getType == DType.STRING => convertStringToDate(cv, options) - case (cv, Some(TimestampType)) if cv.getType == DType.STRING => convertStringToTimestamp(cv, options) }