[SPARK-25127] DataSourceV2: Remove SupportsPushDownCatalystFilters

## What changes were proposed in this pull request?
They depend on internal Expression APIs. Let's see how far we can get without it.

## How was this patch tested?
Just some code removal. There's no existing tests as far as I can tell so it's easy to remove.

Closes #22185 from rxin/SPARK-25127.

Authored-by: Reynold Xin <rxin@databricks.com>
Signed-off-by: Wenchen Fan <wenchen@databricks.com>
This commit is contained in:
Reynold Xin 2018-08-23 08:10:45 +08:00 committed by Wenchen Fan
parent 2bc7b75537
commit 0295ad40de
3 changed files with 0 additions and 66 deletions

View file

@ -1,57 +0,0 @@
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.spark.sql.sources.v2.reader;
import org.apache.spark.annotation.InterfaceStability;
import org.apache.spark.sql.catalyst.expressions.Expression;
/**
* A mix-in interface for {@link ScanConfigBuilder}. Data source readers can implement this
* interface to push down arbitrary expressions as predicates to the data source.
* This is an experimental and unstable interface as {@link Expression} is not public and may get
* changed in the future Spark versions.
*
* Note that, if data source readers implement both this interface and
* {@link SupportsPushDownFilters}, Spark will ignore {@link SupportsPushDownFilters} and only
* process this interface.
*/
@InterfaceStability.Unstable
public interface SupportsPushDownCatalystFilters extends ScanConfigBuilder {
/**
* Pushes down filters, and returns filters that need to be evaluated after scanning.
*/
Expression[] pushCatalystFilters(Expression[] filters);
/**
* Returns the catalyst filters that are pushed to the data source via
* {@link #pushCatalystFilters(Expression[])}.
*
* There are 3 kinds of filters:
* 1. pushable filters which don't need to be evaluated again after scanning.
* 2. pushable filters which still need to be evaluated after scanning, e.g. parquet
* row group filter.
* 3. non-pushable filters.
* Both case 1 and 2 should be considered as pushed filters and should be returned by this method.
*
* It's possible that there is no filters in the query and
* {@link #pushCatalystFilters(Expression[])} is never called, empty array should be returned for
* this case.
*/
Expression[] pushedCatalystFilters();
}

View file

@ -23,10 +23,6 @@ import org.apache.spark.sql.sources.Filter;
/** /**
* A mix-in interface for {@link ScanConfigBuilder}. Data sources can implement this interface to * A mix-in interface for {@link ScanConfigBuilder}. Data sources can implement this interface to
* push down filters to the data source and reduce the size of the data to be read. * push down filters to the data source and reduce the size of the data to be read.
*
* Note that, if data source readers implement both this interface and
* {@link SupportsPushDownCatalystFilters}, Spark will ignore this interface and only process
* {@link SupportsPushDownCatalystFilters}.
*/ */
@InterfaceStability.Evolving @InterfaceStability.Evolving
public interface SupportsPushDownFilters extends ScanConfigBuilder { public interface SupportsPushDownFilters extends ScanConfigBuilder {

View file

@ -40,11 +40,6 @@ object DataSourceV2Strategy extends Strategy {
configBuilder: ScanConfigBuilder, configBuilder: ScanConfigBuilder,
filters: Seq[Expression]): (Seq[Expression], Seq[Expression]) = { filters: Seq[Expression]): (Seq[Expression], Seq[Expression]) = {
configBuilder match { configBuilder match {
case r: SupportsPushDownCatalystFilters =>
val postScanFilters = r.pushCatalystFilters(filters.toArray)
val pushedFilters = r.pushedCatalystFilters()
(pushedFilters, postScanFilters)
case r: SupportsPushDownFilters => case r: SupportsPushDownFilters =>
// A map from translated data source filters to original catalyst filter expressions. // A map from translated data source filters to original catalyst filter expressions.
val translatedFilterToExpr = mutable.HashMap.empty[sources.Filter, Expression] val translatedFilterToExpr = mutable.HashMap.empty[sources.Filter, Expression]