From c175e59f8f665498531b5c4122ad76d22d2cfbb1 Mon Sep 17 00:00:00 2001 From: arnavb Date: Mon, 28 Apr 2025 11:54:37 +0000 Subject: [PATCH] update --- .../columnar/offload/OffloadSingleNodeRules.scala | 9 --------- 1 file changed, 9 deletions(-) diff --git a/gluten-substrait/src/main/scala/org/apache/gluten/extension/columnar/offload/OffloadSingleNodeRules.scala b/gluten-substrait/src/main/scala/org/apache/gluten/extension/columnar/offload/OffloadSingleNodeRules.scala index 2ad6701742a6..d4a82daae5f2 100644 --- a/gluten-substrait/src/main/scala/org/apache/gluten/extension/columnar/offload/OffloadSingleNodeRules.scala +++ b/gluten-substrait/src/main/scala/org/apache/gluten/extension/columnar/offload/OffloadSingleNodeRules.scala @@ -27,7 +27,6 @@ import org.apache.spark.internal.Logging import org.apache.spark.sql.catalyst.optimizer.{BuildLeft, BuildRight, BuildSide} import org.apache.spark.sql.catalyst.plans.logical.Join import org.apache.spark.sql.execution._ -import org.apache.spark.sql.execution.CollectLimitExec import org.apache.spark.sql.execution.RDDScanTransformer import org.apache.spark.sql.execution.aggregate.{HashAggregateExec, ObjectHashAggregateExec, SortAggregateExec} import org.apache.spark.sql.execution.datasources.WriteFilesExec @@ -343,14 +342,6 @@ object OffloadOthers { plan.withReplacement, plan.seed, child) - case plan: CollectLimitExec => - logDebug(s"Columnar Processing for ${plan.getClass} is currently supported.") - val offset = SparkShimLoader.getSparkShims.getCollectLimitOffset(plan) - BackendsApiManager.getSparkPlanExecApiInstance.genColumnarCollectLimitExec( - plan.limit, - plan.child, - offset - ) case plan: RDDScanExec if RDDScanTransformer.isSupportRDDScanExec(plan) => logDebug(s"Columnar Processing for ${plan.getClass} is currently supported.") RDDScanTransformer.getRDDScanTransform(plan)