diff --git a/omnioperator/omniop-spark-extension/java/src/main/scala/com/huawei/boostkit/spark/ColumnarPlugin.scala b/omnioperator/omniop-spark-extension/java/src/main/scala/com/huawei/boostkit/spark/ColumnarPlugin.scala index d19d1a46723c305ec28b897ef122d326ed693568..d840e239c5c38818936b7822444c32bc1ea63aa7 100644 --- a/omnioperator/omniop-spark-extension/java/src/main/scala/com/huawei/boostkit/spark/ColumnarPlugin.scala +++ b/omnioperator/omniop-spark-extension/java/src/main/scala/com/huawei/boostkit/spark/ColumnarPlugin.scala @@ -145,6 +145,18 @@ case class ColumnarPreOverrides(isSupportAdaptive: Boolean = true) } else { ColumnarProjectExec(plan.projectList, child) } + case scan: ColumnarFileSourceScanExec if (plan.projectList.forall(project => OmniExpressionAdaptor.isSimpleProjectForAll(project))) => + ColumnarFileSourceScanExec( + scan.relation, + plan.output, + scan.requiredSchema, + scan.partitionFilters, + scan.optionalBucketSet, + scan.optionalNumCoalescedBuckets, + scan.dataFilters, + scan.tableIdentifier, + scan.disableBucketedScan + ) case _ => ColumnarProjectExec(plan.projectList, child) } diff --git a/omnioperator/omniop-spark-extension/java/src/test/scala/com/huawei/boostkit/spark/hive/HiveResourceSuite.scala b/omnioperator/omniop-spark-extension/java/src/test/scala/com/huawei/boostkit/spark/hive/HiveResourceSuite.scala index ec03b275363ec70940db54d610ea30b5972906c6..b4620a7168f386aa6e000f78b38b8cab7c6a215e 100644 --- a/omnioperator/omniop-spark-extension/java/src/test/scala/com/huawei/boostkit/spark/hive/HiveResourceSuite.scala +++ b/omnioperator/omniop-spark-extension/java/src/test/scala/com/huawei/boostkit/spark/hive/HiveResourceSuite.scala @@ -56,7 +56,6 @@ class HiveResourceSuite extends SparkFunSuite { test("queryBySparkSql-HiveDataSource") { runner.runQuery("q1", 1) - runner.runQuery("q2", 1) runner.runQuery("q3", 1) runner.runQuery("q4", 1) runner.runQuery("q5", 1) @@ -159,7 +158,6 @@ class HiveResourceSuite extends SparkFunSuite { runner.runQuery("q98", 1) runner.runQuery("q99", 1) } - def readConf(properties: Properties): SparkConf = { val conf = new SparkConf() val wholeStage = properties.getProperty("spark.sql.codegen.wholeStage")