spark-instrumented-optimizer/python/pyspark/ml/pipeline.pyi

98 lines
3.7 KiB
Python
Raw Normal View History

#
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from typing import Any, Dict, List, Optional, Tuple, Type, Union
from pyspark.ml._typing import PipelineStage
from pyspark.context import SparkContext
from pyspark.ml.base import Estimator, Model, Transformer
from pyspark.ml.param import Param
from pyspark.ml.util import ( # noqa: F401
DefaultParamsReader as DefaultParamsReader,
DefaultParamsWriter as DefaultParamsWriter,
JavaMLReader as JavaMLReader,
JavaMLWritable as JavaMLWritable,
JavaMLWriter as JavaMLWriter,
MLReadable as MLReadable,
MLReader as MLReader,
MLWritable as MLWritable,
MLWriter as MLWriter,
)
class Pipeline(Estimator[PipelineModel], MLReadable[Pipeline], MLWritable):
stages: List[PipelineStage]
def __init__(self, *, stages: Optional[List[PipelineStage]] = ...) -> None: ...
def setStages(self, stages: List[PipelineStage]) -> Pipeline: ...
def getStages(self) -> List[PipelineStage]: ...
def setParams(self, *, stages: Optional[List[PipelineStage]] = ...) -> Pipeline: ...
def copy(self, extra: Optional[Dict[Param, str]] = ...) -> Pipeline: ...
def write(self) -> JavaMLWriter: ...
def save(self, path: str) -> None: ...
@classmethod
def read(cls) -> PipelineReader: ...
class PipelineWriter(MLWriter):
instance: Pipeline
def __init__(self, instance: Pipeline) -> None: ...
def saveImpl(self, path: str) -> None: ...
class PipelineReader(MLReader[Pipeline]):
cls: Type[Pipeline]
def __init__(self, cls: Type[Pipeline]) -> None: ...
def load(self, path: str) -> Pipeline: ...
class PipelineModelWriter(MLWriter):
instance: PipelineModel
def __init__(self, instance: PipelineModel) -> None: ...
def saveImpl(self, path: str) -> None: ...
class PipelineModelReader(MLReader[PipelineModel]):
cls: Type[PipelineModel]
def __init__(self, cls: Type[PipelineModel]) -> None: ...
def load(self, path: str) -> PipelineModel: ...
class PipelineModel(Model, MLReadable[PipelineModel], MLWritable):
stages: List[PipelineStage]
def __init__(self, stages: List[Transformer]) -> None: ...
def copy(self, extra: Optional[Dict[Param, Any]] = ...) -> PipelineModel: ...
def write(self) -> JavaMLWriter: ...
def save(self, path: str) -> None: ...
@classmethod
def read(cls) -> PipelineModelReader: ...
class PipelineSharedReadWrite:
@staticmethod
def checkStagesForJava(stages: List[PipelineStage]) -> bool: ...
@staticmethod
def validateStages(stages: List[PipelineStage]) -> None: ...
@staticmethod
def saveImpl(
instance: Union[Pipeline, PipelineModel],
stages: List[PipelineStage],
sc: SparkContext,
path: str,
) -> None: ...
@staticmethod
def load(
metadata: Dict[str, Any], sc: SparkContext, path: str
) -> Tuple[str, List[PipelineStage]]: ...
@staticmethod
def getStagePath(
stageUid: str, stageIdx: int, numStages: int, stagesDir: str
) -> str: ...