13fd272cd3
### What changes were proposed in this pull request?
This PR intends to fix typos in the sub-modules:
* `R`
* `common`
* `dev`
* `mlib`
* `external`
* `project`
* `streaming`
* `resource-managers`
* `python`
Split per srowen https://github.com/apache/spark/pull/30323#issuecomment-728981618
NOTE: The misspellings have been reported at 706a726f87 (commitcomment-44064356)
### Why are the changes needed?
Misspelled words make it harder to read / understand content.
### Does this PR introduce _any_ user-facing change?
There are various fixes to documentation, etc...
### How was this patch tested?
No testing was performed
Closes #30402 from jsoref/spelling-R_common_dev_mlib_external_project_streaming_resource-managers_python.
Authored-by: Josh Soref <jsoref@users.noreply.github.com>
Signed-off-by: Sean Owen <srowen@gmail.com>
74 lines
2.5 KiB
Python
74 lines
2.5 KiB
Python
#
|
|
# Licensed to the Apache Software Foundation (ASF) under one
|
|
# or more contributor license agreements. See the NOTICE file
|
|
# distributed with this work for additional information
|
|
# regarding copyright ownership. The ASF licenses this file
|
|
# to you under the Apache License, Version 2.0 (the
|
|
# "License"); you may not use this file except in compliance
|
|
# with the License. You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing,
|
|
# software distributed under the License is distributed on an
|
|
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
# KIND, either express or implied. See the License for the
|
|
# specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
from typing import Callable, Optional, TypeVar
|
|
|
|
from pyspark.accumulators import ( # noqa: F401
|
|
Accumulator as Accumulator,
|
|
AccumulatorParam as AccumulatorParam,
|
|
)
|
|
from pyspark.broadcast import Broadcast as Broadcast # noqa: F401
|
|
from pyspark.conf import SparkConf as SparkConf # noqa: F401
|
|
from pyspark.context import SparkContext as SparkContext # noqa: F401
|
|
from pyspark.files import SparkFiles as SparkFiles # noqa: F401
|
|
from pyspark.status import (
|
|
StatusTracker as StatusTracker,
|
|
SparkJobInfo as SparkJobInfo,
|
|
SparkStageInfo as SparkStageInfo,
|
|
) # noqa: F401
|
|
from pyspark.profiler import ( # noqa: F401
|
|
BasicProfiler as BasicProfiler,
|
|
Profiler as Profiler,
|
|
)
|
|
from pyspark.rdd import RDD as RDD, RDDBarrier as RDDBarrier # noqa: F401
|
|
from pyspark.serializers import ( # noqa: F401
|
|
MarshalSerializer as MarshalSerializer,
|
|
PickleSerializer as PickleSerializer,
|
|
)
|
|
from pyspark.status import ( # noqa: F401
|
|
SparkJobInfo as SparkJobInfo,
|
|
SparkStageInfo as SparkStageInfo,
|
|
StatusTracker as StatusTracker,
|
|
)
|
|
from pyspark.storagelevel import StorageLevel as StorageLevel # noqa: F401
|
|
from pyspark.taskcontext import ( # noqa: F401
|
|
BarrierTaskContext as BarrierTaskContext,
|
|
BarrierTaskInfo as BarrierTaskInfo,
|
|
TaskContext as TaskContext,
|
|
)
|
|
from pyspark.util import InheritableThread as InheritableThread # noqa: F401
|
|
|
|
# Compatibility imports
|
|
from pyspark.sql import ( # noqa: F401
|
|
SQLContext as SQLContext,
|
|
HiveContext as HiveContext,
|
|
Row as Row,
|
|
)
|
|
|
|
T = TypeVar("T")
|
|
F = TypeVar("F", bound=Callable)
|
|
|
|
def since(version: str) -> Callable[[T], T]: ...
|
|
def copy_func(
|
|
f: F,
|
|
name: Optional[str] = ...,
|
|
sinceversion: Optional[str] = ...,
|
|
doc: Optional[str] = ...,
|
|
) -> F: ...
|
|
def keyword_only(func: F) -> F: ...
|