diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/OptimizeSkewedJoin.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/OptimizeSkewedJoin.scala index d3cb8645fc7ea..dd238fe7785bb 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/OptimizeSkewedJoin.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/adaptive/OptimizeSkewedJoin.scala @@ -48,7 +48,7 @@ import org.apache.spark.sql.internal.SQLConf * right: [R1, R2, R3, R4] * * Let's say L2, L4 and R3, R4 are skewed, and each of them get split into 2 sub-partitions. This - * is scheduled to run 4 tasks at the beginning: (L1, R1), (L2, R2), (L2, R2), (L2, R2). + * is scheduled to run 4 tasks at the beginning: (L1, R1), (L2, R2), (L3, R3), (L4, R4). * This rule expands it to 9 tasks to increase parallelism: * (L1, R1), * (L2-1, R2), (L2-2, R2),