From 9d58daa4b90eb90ec7e99a9b28bd51589db18708 Mon Sep 17 00:00:00 2001 From: liurenjie1024 Date: Wed, 10 Apr 2024 17:24:28 +0800 Subject: [PATCH] Test if build works Signed-off-by: liurenjie1024 --- .../org/apache/spark/sql/hive/rapids/GpuHiveTableScanExec.scala | 2 ++ 1 file changed, 2 insertions(+) diff --git a/sql-plugin/src/main/scala/org/apache/spark/sql/hive/rapids/GpuHiveTableScanExec.scala b/sql-plugin/src/main/scala/org/apache/spark/sql/hive/rapids/GpuHiveTableScanExec.scala index be41b4ef82f..f095ad80c81 100644 --- a/sql-plugin/src/main/scala/org/apache/spark/sql/hive/rapids/GpuHiveTableScanExec.scala +++ b/sql-plugin/src/main/scala/org/apache/spark/sql/hive/rapids/GpuHiveTableScanExec.scala @@ -283,6 +283,7 @@ case class GpuHiveTableScanExec(requestedAttributes: Seq[Attribute], val tableLocation: URI = hiveTableRelation.tableMeta.storage.locationUri.getOrElse{ throw new UnsupportedOperationException("Table path not set.") } + println("Hello world!") // No need to check if table directory exists. // FileSystem.listStatus() handles this for GpuHiveTableScanExec, // just like for Apache Spark. @@ -463,6 +464,7 @@ case class GpuHiveTextPartitionReaderFactory(sqlConf: SQLConf, conf, csvOptions, params, partFile, inputFileSchema, requestedOutputDataSchema, maxReaderBatchSizeRows, maxReaderBatchSizeBytes, metrics)) + new AlphabeticallyReorderingColumnPartitionReader(reader, partFile.partitionValues, partitionSchema,