[SPARK-35378][SQL][FOLLOWUP] Move CommandResult to catalyst.plans.logical
### What changes were proposed in this pull request? https://github.com/apache/spark/pull/32513 added the case class `CommandResult` in package `org.apache.spark.sql.expression`. It is not suitable, so this PR move `CommandResult` from `org.apache.spark.sql.expression` to `org.apache.spark.sql.catalyst.plans.logical`. ### Why are the changes needed? Make `CommandResult` in suitable package. ### Does this PR introduce _any_ user-facing change? 'No'. ### How was this patch tested? No need. Closes #32942 from beliefer/SPARK-35378-followup. Lead-authored-by: gengjiaan <gengjiaan@360.cn> Co-authored-by: Jiaan Geng <beliefer@163.com> Signed-off-by: Dongjoon Hyun <dongjoon@apache.org>
This commit is contained in:
parent
abf9675a75
commit
ee2d8ae322
|
@ -15,12 +15,11 @@
|
|||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.spark.sql.expressions
|
||||
package org.apache.spark.sql.catalyst.plans.logical
|
||||
|
||||
import org.apache.spark.sql.catalyst.InternalRow
|
||||
import org.apache.spark.sql.catalyst.expressions.Attribute
|
||||
import org.apache.spark.sql.catalyst.plans.QueryPlan
|
||||
import org.apache.spark.sql.catalyst.plans.logical.{LeafNode, LogicalPlan, Statistics}
|
||||
import org.apache.spark.sql.catalyst.plans.logical.statsEstimation.EstimationUtils
|
||||
import org.apache.spark.sql.execution.SparkPlan
|
||||
|
|
@ -30,7 +30,7 @@ import org.apache.spark.sql.catalyst.{InternalRow, QueryPlanningTracker}
|
|||
import org.apache.spark.sql.catalyst.analysis.UnsupportedOperationChecker
|
||||
import org.apache.spark.sql.catalyst.expressions.codegen.ByteCodeStats
|
||||
import org.apache.spark.sql.catalyst.plans.QueryPlan
|
||||
import org.apache.spark.sql.catalyst.plans.logical.{AppendData, Command, CreateTableAsSelect, LogicalPlan, OverwriteByExpression, OverwritePartitionsDynamic, ReplaceTableAsSelect, ReturnAnswer}
|
||||
import org.apache.spark.sql.catalyst.plans.logical.{AppendData, Command, CommandResult, CreateTableAsSelect, LogicalPlan, OverwriteByExpression, OverwritePartitionsDynamic, ReplaceTableAsSelect, ReturnAnswer}
|
||||
import org.apache.spark.sql.catalyst.rules.{PlanChangeLogger, Rule}
|
||||
import org.apache.spark.sql.catalyst.util.StringUtils.PlanStringConcat
|
||||
import org.apache.spark.sql.catalyst.util.truncatedString
|
||||
|
@ -39,7 +39,6 @@ import org.apache.spark.sql.execution.bucketing.{CoalesceBucketsInJoin, DisableU
|
|||
import org.apache.spark.sql.execution.dynamicpruning.PlanDynamicPruningFilters
|
||||
import org.apache.spark.sql.execution.exchange.{EnsureRequirements, ReuseExchange}
|
||||
import org.apache.spark.sql.execution.streaming.{IncrementalExecution, OffsetSeqMetadata}
|
||||
import org.apache.spark.sql.expressions.CommandResult
|
||||
import org.apache.spark.sql.internal.SQLConf
|
||||
import org.apache.spark.sql.streaming.OutputMode
|
||||
import org.apache.spark.util.Utils
|
||||
|
|
|
@ -36,7 +36,6 @@ import org.apache.spark.sql.execution.exchange.{REPARTITION_BY_COL, REPARTITION_
|
|||
import org.apache.spark.sql.execution.python._
|
||||
import org.apache.spark.sql.execution.streaming._
|
||||
import org.apache.spark.sql.execution.streaming.sources.MemoryPlan
|
||||
import org.apache.spark.sql.expressions.CommandResult
|
||||
import org.apache.spark.sql.internal.SQLConf
|
||||
import org.apache.spark.sql.streaming.OutputMode
|
||||
import org.apache.spark.sql.types.StructType
|
||||
|
|
|
@ -21,10 +21,9 @@ import scala.io.Source
|
|||
import org.apache.spark.sql.{AnalysisException, FastOperator}
|
||||
import org.apache.spark.sql.catalyst.analysis.UnresolvedNamespace
|
||||
import org.apache.spark.sql.catalyst.plans.QueryPlan
|
||||
import org.apache.spark.sql.catalyst.plans.logical.{LogicalPlan, OneRowRelation, Project, ShowTables, SubqueryAlias}
|
||||
import org.apache.spark.sql.catalyst.plans.logical.{CommandResult, LogicalPlan, OneRowRelation, Project, ShowTables, SubqueryAlias}
|
||||
import org.apache.spark.sql.catalyst.trees.TreeNodeTag
|
||||
import org.apache.spark.sql.execution.command.{ExecutedCommandExec, ShowTablesCommand}
|
||||
import org.apache.spark.sql.expressions.CommandResult
|
||||
import org.apache.spark.sql.internal.SQLConf
|
||||
import org.apache.spark.sql.test.SharedSparkSession
|
||||
import org.apache.spark.util.Utils
|
||||
|
|
Loading…
Reference in a new issue