spark-instrumented-optimizer/sql/core/src/main
Cheng Hao 1baaf2b9bd [SPARK-10829] [SQL] Filter combine partition key and attribute doesn't work in DataSource scan
```scala
withSQLConf(SQLConf.PARQUET_FILTER_PUSHDOWN_ENABLED.key -> "true") {
      withTempPath { dir =>
        val path = s"${dir.getCanonicalPath}/part=1"
        (1 to 3).map(i => (i, i.toString)).toDF("a", "b").write.parquet(path)

        // If the "part = 1" filter gets pushed down, this query will throw an exception since
        // "part" is not a valid column in the actual Parquet file
        checkAnswer(
          sqlContext.read.parquet(path).filter("a > 0 and (part = 0 or a > 1)"),
          (2 to 3).map(i => Row(i, i.toString, 1)))
      }
    }
```

We expect the result to be:
```
2,1
3,1
```
But got
```
1,1
2,1
3,1
```

Author: Cheng Hao <hao.cheng@intel.com>

Closes #8916 from chenghao-intel/partition_filter.
2015-10-14 16:29:32 -07:00
..
java/org/apache/spark/sql [SPARK-10474] [SQL] Aggregation fails to allocate memory for pointer array (round 2) 2015-09-23 19:34:31 -07:00
resources [SPARK-9763][SQL] Minimize exposure of internal SQL classes. 2015-08-10 13:49:23 -07:00
scala/org/apache/spark/sql [SPARK-10829] [SQL] Filter combine partition key and attribute doesn't work in DataSource scan 2015-10-14 16:29:32 -07:00