4d2b559d92
### What changes were proposed in this pull request? Consolidate PySpark testing utils by removing `python/pyspark/pandas/testing`, and then creating a file `pandasutils` under `python/pyspark/testing` for test utilities used in `pyspark/pandas`. ### Why are the changes needed? `python/pyspark/pandas/testing` hold test utilites for pandas-on-spark, and `python/pyspark/testing` contain test utilities for pyspark. Consolidating them makes code cleaner and easier to maintain. Updated import statements are as shown below: - from pyspark.testing.sqlutils import SQLTestUtils - from pyspark.testing.pandasutils import PandasOnSparkTestCase, TestUtils (PandasOnSparkTestCase is the original ReusedSQLTestCase in `python/pyspark/pandas/testing/utils.py`) Minor improvements include: - Usage of missing library's requirement_message - `except ImportError` rather than `except` - import pyspark.pandas alias as `ps` rather than `pp` ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? Unit tests under python/pyspark/pandas/tests. Closes #32177 from xinrong-databricks/port.merge_utils. Authored-by: Xinrong Meng <xinrong.meng@databricks.com> Signed-off-by: Takuya UESHIN <ueshin@databricks.com>
102 lines
4 KiB
Python
102 lines
4 KiB
Python
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
# this work for additional information regarding copyright ownership.
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
import unittest
|
|
|
|
import pandas as pd
|
|
import numpy as np
|
|
|
|
from pyspark import pandas as ps
|
|
from pyspark.pandas.plot import PandasOnSparkPlotAccessor, BoxPlotBase
|
|
from pyspark.testing.pandasutils import have_plotly, plotly_requirement_message
|
|
|
|
|
|
class SeriesPlotTest(unittest.TestCase):
|
|
@property
|
|
def pdf1(self):
|
|
return pd.DataFrame(
|
|
{"a": [1, 2, 3, 4, 5, 6, 7, 8, 9, 15, 50]}, index=[0, 1, 3, 5, 6, 8, 9, 9, 9, 10, 10]
|
|
)
|
|
|
|
@property
|
|
def kdf1(self):
|
|
return ps.from_pandas(self.pdf1)
|
|
|
|
@unittest.skipIf(not have_plotly, plotly_requirement_message)
|
|
def test_plot_backends(self):
|
|
plot_backend = "plotly"
|
|
|
|
with ps.option_context("plotting.backend", plot_backend):
|
|
self.assertEqual(ps.options.plotting.backend, plot_backend)
|
|
|
|
module = PandasOnSparkPlotAccessor._get_plot_backend(plot_backend)
|
|
self.assertEqual(module.__name__, "pyspark.pandas.plot.plotly")
|
|
|
|
def test_plot_backends_incorrect(self):
|
|
fake_plot_backend = "none_plotting_module"
|
|
|
|
with ps.option_context("plotting.backend", fake_plot_backend):
|
|
self.assertEqual(ps.options.plotting.backend, fake_plot_backend)
|
|
|
|
with self.assertRaises(ValueError):
|
|
PandasOnSparkPlotAccessor._get_plot_backend(fake_plot_backend)
|
|
|
|
def test_box_summary(self):
|
|
def check_box_summary(kdf, pdf):
|
|
k = 1.5
|
|
stats, fences = BoxPlotBase.compute_stats(kdf["a"], "a", whis=k, precision=0.01)
|
|
outliers = BoxPlotBase.outliers(kdf["a"], "a", *fences)
|
|
whiskers = BoxPlotBase.calc_whiskers("a", outliers)
|
|
fliers = BoxPlotBase.get_fliers("a", outliers, whiskers[0])
|
|
|
|
expected_mean = pdf["a"].mean()
|
|
expected_median = pdf["a"].median()
|
|
expected_q1 = np.percentile(pdf["a"], 25)
|
|
expected_q3 = np.percentile(pdf["a"], 75)
|
|
iqr = expected_q3 - expected_q1
|
|
expected_fences = (expected_q1 - k * iqr, expected_q3 + k * iqr)
|
|
pdf["outlier"] = ~pdf["a"].between(fences[0], fences[1])
|
|
expected_whiskers = (
|
|
pdf.query("not outlier")["a"].min(),
|
|
pdf.query("not outlier")["a"].max(),
|
|
)
|
|
expected_fliers = pdf.query("outlier")["a"].values
|
|
|
|
self.assertEqual(expected_mean, stats["mean"])
|
|
self.assertEqual(expected_median, stats["med"])
|
|
self.assertEqual(expected_q1, stats["q1"] + 0.5)
|
|
self.assertEqual(expected_q3, stats["q3"] - 0.5)
|
|
self.assertEqual(expected_fences[0], fences[0] + 2.0)
|
|
self.assertEqual(expected_fences[1], fences[1] - 2.0)
|
|
self.assertEqual(expected_whiskers[0], whiskers[0])
|
|
self.assertEqual(expected_whiskers[1], whiskers[1])
|
|
self.assertEqual(expected_fliers, fliers)
|
|
|
|
check_box_summary(self.kdf1, self.pdf1)
|
|
check_box_summary(-self.kdf1, -self.pdf1)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
from pyspark.pandas.tests.plot.test_series_plot import * # noqa: F401
|
|
|
|
try:
|
|
import xmlrunner # type: ignore[import]
|
|
testRunner = xmlrunner.XMLTestRunner(output='target/test-reports', verbosity=2)
|
|
except ImportError:
|
|
testRunner = None
|
|
unittest.main(testRunner=testRunner, verbosity=2)
|