| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """The `Step` definitions for SageMaker Pipelines Workflows.""" |
| | from __future__ import absolute_import |
| |
|
| | import abc |
| | import warnings |
| |
|
| | from enum import Enum |
| | from typing import Dict, List, Set, Union, Optional, Any, TYPE_CHECKING |
| | from urllib.parse import urlparse |
| |
|
| | import attr |
| |
|
| | from sagemaker import Session |
| | from sagemaker.estimator import EstimatorBase, _TrainingJob |
| | from sagemaker.inputs import CreateModelInput, TrainingInput, TransformInput, FileSystemInput |
| | from sagemaker.model import Model |
| | from sagemaker.pipeline import PipelineModel |
| | from sagemaker.processing import ( |
| | ProcessingInput, |
| | ProcessingJob, |
| | ProcessingOutput, |
| | Processor, |
| | ) |
| | from sagemaker.transformer import Transformer, _TransformJob |
| | from sagemaker.tuner import HyperparameterTuner, _TuningJob |
| | from sagemaker.workflow.conditions import Condition |
| | from sagemaker.workflow import is_pipeline_variable |
| | from sagemaker.workflow.entities import ( |
| | DefaultEnumMeta, |
| | Entity, |
| | RequestType, |
| | ) |
| | from sagemaker.workflow.pipeline_context import _JobStepArguments |
| | from sagemaker.workflow.properties import ( |
| | PropertyFile, |
| | Properties, |
| | ) |
| | from sagemaker.workflow.entities import PipelineVariable |
| | from sagemaker.workflow.functions import Join, JsonGet |
| | from sagemaker.workflow.retry import RetryPolicy |
| |
|
| | if TYPE_CHECKING: |
| | from sagemaker.workflow.step_collections import StepCollection |
| |
|
| |
|
| | class StepTypeEnum(Enum, metaclass=DefaultEnumMeta): |
| | """Enum of `Step` types.""" |
| |
|
| | CONDITION = "Condition" |
| | CREATE_MODEL = "Model" |
| | PROCESSING = "Processing" |
| | REGISTER_MODEL = "RegisterModel" |
| | TRAINING = "Training" |
| | TRANSFORM = "Transform" |
| | CALLBACK = "Callback" |
| | TUNING = "Tuning" |
| | LAMBDA = "Lambda" |
| | QUALITY_CHECK = "QualityCheck" |
| | CLARIFY_CHECK = "ClarifyCheck" |
| | EMR = "EMR" |
| | FAIL = "Fail" |
| |
|
| |
|
| | @attr.s |
| | class Step(Entity): |
| | """Pipeline `Step` for SageMaker Pipelines Workflows. |
| | |
| | Attributes: |
| | name (str): The name of the `Step`. |
| | display_name (str): The display name of the `Step`. |
| | description (str): The description of the `Step`. |
| | step_type (StepTypeEnum): The type of the `Step`. |
| | depends_on (List[Union[str, Step, StepCollection]]): The list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that the current `Step` |
| | depends on. |
| | """ |
| |
|
| | name: str = attr.ib(factory=str) |
| | display_name: Optional[str] = attr.ib(default=None) |
| | description: Optional[str] = attr.ib(default=None) |
| | step_type: StepTypeEnum = attr.ib(factory=StepTypeEnum.factory) |
| | depends_on: Optional[List[Union[str, "Step", "StepCollection"]]] = attr.ib(default=None) |
| |
|
| | @property |
| | @abc.abstractmethod |
| | def arguments(self) -> RequestType: |
| | """The arguments to the particular `Step` service call.""" |
| |
|
| | @property |
| | def step_only_arguments(self) -> RequestType: |
| | """The arguments to this Step only. |
| | |
| | Compound Steps such as the ConditionStep will have to |
| | override this method to return arguments pertaining to only that step. |
| | """ |
| | return self.arguments |
| |
|
| | @property |
| | @abc.abstractmethod |
| | def properties(self): |
| | """The properties of the particular `Step`.""" |
| |
|
| | def to_request(self) -> RequestType: |
| | """Gets the request structure for workflow service calls.""" |
| | request_dict = { |
| | "Name": self.name, |
| | "Type": self.step_type.value, |
| | "Arguments": self.arguments, |
| | } |
| | if self.depends_on: |
| | request_dict["DependsOn"] = self._resolve_depends_on(self.depends_on) |
| | if self.display_name: |
| | request_dict["DisplayName"] = self.display_name |
| | if self.description: |
| | request_dict["Description"] = self.description |
| |
|
| | return request_dict |
| |
|
| | def add_depends_on(self, step_names: List[Union[str, "Step", "StepCollection"]]): |
| | """Add `Step` names or `Step` instances to the current `Step` depends on list.""" |
| |
|
| | if not step_names: |
| | return |
| |
|
| | if not self.depends_on: |
| | self.depends_on = [] |
| | self.depends_on.extend(step_names) |
| |
|
| | @property |
| | def ref(self) -> Dict[str, str]: |
| | """Gets a reference dictionary for `Step` instances.""" |
| | return {"Name": self.name} |
| |
|
| | @staticmethod |
| | def _resolve_depends_on( |
| | depends_on_list: List[Union[str, "Step", "StepCollection"]] |
| | ) -> List[str]: |
| | """Resolve the `Step` depends on list.""" |
| | from sagemaker.workflow.step_collections import StepCollection |
| |
|
| | depends_on = [] |
| | for step in depends_on_list: |
| | |
| | |
| | if isinstance(step, (Step, StepCollection)): |
| | depends_on.append(step.name) |
| | elif isinstance(step, str): |
| | depends_on.append(step) |
| | else: |
| | raise ValueError(f"Invalid input step type: {type(step)}") |
| | return depends_on |
| |
|
| | def _find_step_dependencies( |
| | self, step_map: Dict[str, Union["Step", "StepCollection"]] |
| | ) -> List[str]: |
| | """Find the all step names this step is dependent on.""" |
| | step_dependencies = set() |
| | if self.depends_on: |
| | step_dependencies.update(self._find_dependencies_in_depends_on_list(step_map)) |
| | step_dependencies.update( |
| | self._find_dependencies_in_step_arguments(self.step_only_arguments, step_map) |
| | ) |
| | return list(step_dependencies) |
| |
|
| | def _find_dependencies_in_depends_on_list( |
| | self, step_map: Dict[str, Union["Step", "StepCollection"]] |
| | ) -> Set[str]: |
| | """Find dependency steps referenced in the depends-on field of this step.""" |
| | |
| | from sagemaker.workflow.step_collections import StepCollection |
| |
|
| | dependencies = set() |
| | for step in self.depends_on: |
| | if isinstance(step, Step): |
| | dependencies.add(step.name) |
| | elif isinstance(step, StepCollection): |
| | dependencies.add(step.steps[-1].name) |
| | elif isinstance(step, str): |
| | |
| | dependencies.add(self._get_step_name_from_str(step, step_map)) |
| | return dependencies |
| |
|
| | def _find_dependencies_in_step_arguments( |
| | self, obj: Any, step_map: Dict[str, Union["Step", "StepCollection"]] |
| | ): |
| | """Find the step dependencies referenced in the arguments of this step.""" |
| | dependencies = set() |
| | if isinstance(obj, dict): |
| | for value in obj.values(): |
| | if isinstance(value, (PipelineVariable, Condition)): |
| | for referenced_step in value._referenced_steps: |
| | dependencies.add(self._get_step_name_from_str(referenced_step, step_map)) |
| | if isinstance(value, JsonGet): |
| | self._validate_json_get_function(value, step_map) |
| | dependencies.update(self._find_dependencies_in_step_arguments(value, step_map)) |
| | elif isinstance(obj, list): |
| | for item in obj: |
| | if isinstance(item, (PipelineVariable, Condition)): |
| | for referenced_step in item._referenced_steps: |
| | dependencies.add(self._get_step_name_from_str(referenced_step, step_map)) |
| | if isinstance(item, JsonGet): |
| | self._validate_json_get_function(item, step_map) |
| | dependencies.update(self._find_dependencies_in_step_arguments(item, step_map)) |
| | return dependencies |
| |
|
| | def _validate_json_get_function( |
| | self, json_get: JsonGet, step_map: Dict[str, Union["Step", "StepCollection"]] |
| | ): |
| | """Validate the JsonGet function inputs.""" |
| | property_file_reference = json_get.property_file |
| | processing_step = step_map[json_get.step_name] |
| | property_file = None |
| | if isinstance(property_file_reference, str): |
| | if not isinstance(processing_step, ProcessingStep): |
| | raise ValueError( |
| | f"Invalid JsonGet function {json_get.expr} in step '{self.name}'. JsonGet " |
| | f"function can only be evaluated on processing step outputs." |
| | ) |
| | for file in processing_step.property_files: |
| | if file.name == property_file_reference: |
| | property_file = file |
| | break |
| | elif isinstance(property_file_reference, PropertyFile): |
| | property_file = property_file_reference |
| | if property_file is None: |
| | raise ValueError( |
| | f"Invalid JsonGet function {json_get.expr} in step '{self.name}'. Property file " |
| | f"reference '{property_file_reference}' is undefined in step " |
| | f"'{processing_step.name}'." |
| | ) |
| | property_file_output = None |
| | if "ProcessingOutputConfig" in processing_step.arguments: |
| | for output in processing_step.arguments["ProcessingOutputConfig"]["Outputs"]: |
| | if output["OutputName"] == property_file.output_name: |
| | property_file_output = output |
| | if property_file_output is None: |
| | raise ValueError( |
| | f"Processing output name '{property_file.output_name}' defined in property file " |
| | f"'{property_file.name}' not found in processing step '{processing_step.name}'." |
| | ) |
| |
|
| | @staticmethod |
| | def _get_step_name_from_str( |
| | str_input: str, step_map: Dict[str, Union["Step", "StepCollection"]] |
| | ) -> str: |
| | """Convert a Step or StepCollection name input to step name.""" |
| | from sagemaker.workflow.step_collections import StepCollection |
| |
|
| | if str_input not in step_map: |
| | raise ValueError(f"Step {str_input} is undefined.") |
| | if isinstance(step_map[str_input], StepCollection): |
| | return step_map[str_input].steps[-1].name |
| | return str_input |
| |
|
| | @staticmethod |
| | def _trim_experiment_config(request_dict: Dict): |
| | """For job steps, trim the experiment config to keep the trial component display name.""" |
| | if request_dict.get("ExperimentConfig", {}).get("TrialComponentDisplayName"): |
| | request_dict["ExperimentConfig"] = { |
| | "TrialComponentDisplayName": request_dict["ExperimentConfig"][ |
| | "TrialComponentDisplayName" |
| | ] |
| | } |
| | else: |
| | request_dict.pop("ExperimentConfig", None) |
| |
|
| |
|
| | @attr.s |
| | class CacheConfig: |
| | """Configuration class to enable caching in SageMaker Pipelines Workflows. |
| | |
| | If caching is enabled, the pipeline attempts to find a previous execution of a `Step` |
| | that was called with the same arguments. `Step` caching only considers successful execution. |
| | If a successful previous execution is found, the pipeline propagates the values |
| | from the previous execution rather than recomputing the `Step`. |
| | When multiple successful executions exist within the timeout period, |
| | it uses the result for the most recent successful execution. |
| | |
| | |
| | Attributes: |
| | enable_caching (bool): To enable `Step` caching. Defaults to `False`. |
| | expire_after (str): If `Step` caching is enabled, a timeout also needs to defined. |
| | It defines how old a previous execution can be to be considered for reuse. |
| | Value should be an ISO 8601 duration string. Defaults to `None`. |
| | |
| | Examples:: |
| | |
| | 'p30d' # 30 days |
| | 'P4DT12H' # 4 days and 12 hours |
| | 'T12H' # 12 hours |
| | """ |
| |
|
| | enable_caching: bool = attr.ib(default=False) |
| | expire_after = attr.ib( |
| | default=None, validator=attr.validators.optional(attr.validators.instance_of(str)) |
| | ) |
| |
|
| | @property |
| | def config(self): |
| | """Configures `Step` caching for SageMaker Pipelines Workflows.""" |
| | config = {"Enabled": self.enable_caching} |
| | if self.expire_after is not None: |
| | config["ExpireAfter"] = self.expire_after |
| | return {"CacheConfig": config} |
| |
|
| |
|
| | class ConfigurableRetryStep(Step): |
| | """`ConfigurableRetryStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_type: StepTypeEnum, |
| | display_name: str = None, |
| | description: str = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: List[RetryPolicy] = None, |
| | ): |
| | super().__init__( |
| | name=name, |
| | display_name=display_name, |
| | step_type=step_type, |
| | description=description, |
| | depends_on=depends_on, |
| | ) |
| | self.retry_policies = [] if not retry_policies else retry_policies |
| |
|
| | def add_retry_policy(self, retry_policy: RetryPolicy): |
| | """Add a policy to the current `ConfigurableRetryStep` retry policies list.""" |
| | if not retry_policy: |
| | return |
| |
|
| | if not self.retry_policies: |
| | self.retry_policies = [] |
| | self.retry_policies.append(retry_policy) |
| |
|
| | def to_request(self) -> RequestType: |
| | """Gets the request structure for `ConfigurableRetryStep`.""" |
| | step_dict = super().to_request() |
| | if self.retry_policies: |
| | step_dict["RetryPolicies"] = self._resolve_retry_policy(self.retry_policies) |
| | return step_dict |
| |
|
| | @staticmethod |
| | def _resolve_retry_policy(retry_policy_list: List[RetryPolicy]) -> List[RequestType]: |
| | """Resolve the `ConfigurableRetryStep` retry policy list.""" |
| | return [retry_policy.to_request() for retry_policy in retry_policy_list] |
| |
|
| |
|
| | class TrainingStep(ConfigurableRetryStep): |
| | """`TrainingStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_args: _JobStepArguments = None, |
| | estimator: EstimatorBase = None, |
| | display_name: str = None, |
| | description: str = None, |
| | inputs: Union[TrainingInput, dict, str, FileSystemInput] = None, |
| | cache_config: CacheConfig = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: List[RetryPolicy] = None, |
| | ): |
| | """Construct a `TrainingStep`, given an `EstimatorBase` instance. |
| | |
| | In addition to the `EstimatorBase` instance, the other arguments are those |
| | that are supplied to the `fit` method of the `sagemaker.estimator.Estimator`. |
| | |
| | Args: |
| | name (str): The name of the `TrainingStep`. |
| | step_args (_JobStepArguments): The arguments for the `TrainingStep` definition. |
| | estimator (EstimatorBase): A `sagemaker.estimator.EstimatorBase` instance. |
| | display_name (str): The display name of the `TrainingStep`. |
| | description (str): The description of the `TrainingStep`. |
| | inputs (Union[str, dict, TrainingInput, FileSystemInput]): Information |
| | about the training data. This can be one of three types: |
| | |
| | * (str) the S3 location where training data is saved, or a file:// path in |
| | local mode. |
| | * (dict[str, str] or dict[str, sagemaker.inputs.TrainingInput]) If using multiple |
| | channels for training data, you can specify a dictionary mapping channel names to |
| | strings or :func:`~sagemaker.inputs.TrainingInput` objects. |
| | * (sagemaker.inputs.TrainingInput) - channel configuration for S3 data sources |
| | that can provide additional information as well as the path to the training |
| | dataset. |
| | See :func:`sagemaker.inputs.TrainingInput` for full details. |
| | * (sagemaker.inputs.FileSystemInput) - channel configuration for |
| | a file system data source that can provide additional information as well as |
| | the path to the training dataset. |
| | |
| | cache_config (CacheConfig): A `sagemaker.workflow.steps.CacheConfig` instance. |
| | depends_on (List[Union[str, Step, StepCollection]]): A list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that this `TrainingStep` |
| | depends on. |
| | retry_policies (List[RetryPolicy]): A list of retry policies. |
| | """ |
| | super(TrainingStep, self).__init__( |
| | name, StepTypeEnum.TRAINING, display_name, description, depends_on, retry_policies |
| | ) |
| |
|
| | if not (step_args is not None) ^ (estimator is not None): |
| | raise ValueError("Either step_args or estimator need to be given.") |
| |
|
| | if step_args: |
| | from sagemaker.workflow.utilities import validate_step_args_input |
| |
|
| | validate_step_args_input( |
| | step_args=step_args, |
| | expected_caller={Session.train.__name__}, |
| | error_message="The step_args of TrainingStep must be obtained from estimator.fit().", |
| | ) |
| |
|
| | self.step_args = step_args.args if step_args else None |
| | self.estimator = estimator |
| | self.inputs = inputs |
| |
|
| | self._properties = Properties(step_name=name, shape_name="DescribeTrainingJobResponse") |
| | self.cache_config = cache_config |
| |
|
| | if self.cache_config: |
| | if (self.step_args and "ProfilerConfig" in self.step_args) or ( |
| | self.estimator is not None and not self.estimator.disable_profiler |
| | ): |
| | msg = ( |
| | "Profiling is enabled on the provided estimator. " |
| | "The default profiler rule includes a timestamp " |
| | "which will change each time the pipeline is " |
| | "upserted, causing cache misses. If profiling " |
| | "is not needed, set disable_profiler to True on the estimator." |
| | ) |
| | warnings.warn(msg) |
| |
|
| | if not self.step_args: |
| | warnings.warn( |
| | ( |
| | 'We are deprecating the instantiation of TrainingStep using "estimator".' |
| | 'Instead, simply using "step_args".' |
| | ), |
| | DeprecationWarning, |
| | ) |
| |
|
| | self.job_name = None |
| | if estimator and (estimator.source_dir or estimator.entry_point): |
| | |
| | |
| | |
| | |
| | |
| | |
| | if not is_pipeline_variable(estimator.source_dir) and not is_pipeline_variable( |
| | estimator.entry_point |
| | ): |
| | self.job_name = self._generate_code_upload_path() |
| |
|
| | @property |
| | def arguments(self) -> RequestType: |
| | """The arguments dictionary that is used to call `create_training_job`. |
| | |
| | NOTE: The `CreateTrainingJob` request is not quite the args list that workflow needs. |
| | The `TrainingJobName` and `ExperimentConfig` attributes cannot be included. |
| | """ |
| | if self.step_args: |
| | request_dict = self.step_args |
| | else: |
| | self.estimator._prepare_for_training(self.job_name) |
| | train_args = _TrainingJob._get_train_args( |
| | self.estimator, self.inputs, experiment_config=dict() |
| | ) |
| | request_dict = self.estimator.sagemaker_session._get_train_request(**train_args) |
| |
|
| | if "HyperParameters" in request_dict: |
| | request_dict["HyperParameters"].pop("sagemaker_job_name", None) |
| |
|
| | request_dict.pop("TrainingJobName", None) |
| | Step._trim_experiment_config(request_dict) |
| |
|
| | return request_dict |
| |
|
| | @property |
| | def properties(self): |
| | """A `Properties` object representing the `DescribeTrainingJobResponse` data model.""" |
| | return self._properties |
| |
|
| | def to_request(self) -> RequestType: |
| | """Updates the request dictionary with cache configuration.""" |
| | request_dict = super().to_request() |
| | if self.cache_config: |
| | request_dict.update(self.cache_config.config) |
| |
|
| | return request_dict |
| |
|
| | def _generate_code_upload_path(self) -> str or None: |
| | """Generate an upload path for local training scripts based on their content.""" |
| | from sagemaker.workflow.utilities import hash_files_or_dirs |
| |
|
| | if self.estimator.source_dir: |
| | source_dir_url = urlparse(self.estimator.source_dir) |
| | if source_dir_url.scheme == "" or source_dir_url.scheme == "file": |
| | code_hash = hash_files_or_dirs( |
| | [self.estimator.source_dir] + self.estimator.dependencies |
| | ) |
| | return f"{self.name}-{code_hash}"[:1024] |
| | elif self.estimator.entry_point: |
| | entry_point_url = urlparse(self.estimator.entry_point) |
| | if entry_point_url.scheme == "" or entry_point_url.scheme == "file": |
| | code_hash = hash_files_or_dirs( |
| | [self.estimator.entry_point] + self.estimator.dependencies |
| | ) |
| | return f"{self.name}-{code_hash}"[:1024] |
| | return None |
| |
|
| |
|
| | class CreateModelStep(ConfigurableRetryStep): |
| | """`CreateModelStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_args: Optional[dict] = None, |
| | model: Optional[Union[Model, PipelineModel]] = None, |
| | inputs: Optional[CreateModelInput] = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: Optional[List[RetryPolicy]] = None, |
| | display_name: Optional[str] = None, |
| | description: Optional[str] = None, |
| | ): |
| | """Construct a `CreateModelStep`, given an `sagemaker.model.Model` instance. |
| | |
| | In addition to the `Model` instance, the other arguments are those that are supplied to |
| | the `_create_sagemaker_model` method of the `sagemaker.model.Model._create_sagemaker_model`. |
| | |
| | Args: |
| | name (str): The name of the `CreateModelStep`. |
| | step_args (dict): The arguments for the `CreateModelStep` definition (default: None). |
| | model (Model or PipelineModel): A `sagemaker.model.Model` |
| | or `sagemaker.pipeline.PipelineModel` instance (default: None). |
| | inputs (CreateModelInput): A `sagemaker.inputs.CreateModelInput` instance. |
| | (default: None). |
| | depends_on (List[Union[str, Step, StepCollection]]): A list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that this `CreateModelStep` |
| | depends on (default: None). |
| | retry_policies (List[RetryPolicy]): A list of retry policies (default: None). |
| | display_name (str): The display name of the `CreateModelStep` (default: None). |
| | description (str): The description of the `CreateModelStep` (default: None). |
| | """ |
| | super(CreateModelStep, self).__init__( |
| | name, StepTypeEnum.CREATE_MODEL, display_name, description, depends_on, retry_policies |
| | ) |
| | if not (step_args is None) ^ (model is None): |
| | raise ValueError( |
| | "step_args and model are mutually exclusive. Either of them should be provided." |
| | ) |
| |
|
| | self.step_args = step_args |
| | self.model = model |
| | self.inputs = inputs or CreateModelInput() |
| |
|
| | self._properties = Properties(step_name=name, shape_name="DescribeModelOutput") |
| |
|
| | warnings.warn( |
| | ( |
| | "We are deprecating the use of CreateModelStep. " |
| | "Instead, please use the ModelStep, which simply takes in the step arguments " |
| | "generated by model.create(). For more, see: " |
| | "https://sagemaker.readthedocs.io/en/stable/" |
| | "amazon_sagemaker_model_building_pipeline.html#model-step" |
| | ), |
| | DeprecationWarning, |
| | ) |
| |
|
| | @property |
| | def arguments(self) -> RequestType: |
| | """The arguments dictionary that is used to call `create_model`. |
| | |
| | NOTE: The `CreateModelRequest` is not quite the args list that workflow needs. |
| | `ModelName` cannot be included in the arguments. |
| | """ |
| |
|
| | if self.step_args: |
| | request_dict = self.step_args |
| | else: |
| | if isinstance(self.model, PipelineModel): |
| | request_dict = self.model.sagemaker_session._create_model_request( |
| | name="", |
| | role=self.model.role, |
| | container_defs=self.model.pipeline_container_def(self.inputs.instance_type), |
| | vpc_config=self.model.vpc_config, |
| | enable_network_isolation=self.model.enable_network_isolation, |
| | ) |
| | else: |
| | request_dict = self.model.sagemaker_session._create_model_request( |
| | name="", |
| | role=self.model.role, |
| | container_defs=self.model.prepare_container_def( |
| | instance_type=self.inputs.instance_type, |
| | accelerator_type=self.inputs.accelerator_type, |
| | ), |
| | vpc_config=self.model.vpc_config, |
| | enable_network_isolation=self.model.enable_network_isolation(), |
| | ) |
| | request_dict.pop("ModelName", None) |
| |
|
| | return request_dict |
| |
|
| | @property |
| | def properties(self): |
| | """A `Properties` object representing the `DescribeModelResponse` data model.""" |
| | return self._properties |
| |
|
| |
|
| | class TransformStep(ConfigurableRetryStep): |
| | """`TransformStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_args: _JobStepArguments = None, |
| | transformer: Transformer = None, |
| | inputs: TransformInput = None, |
| | display_name: str = None, |
| | description: str = None, |
| | cache_config: CacheConfig = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: List[RetryPolicy] = None, |
| | ): |
| | """Constructs a `TransformStep`, given a `Transformer` instance. |
| | |
| | In addition to the `Transformer` instance, the other arguments are those |
| | that are supplied to the `transform` method of the `sagemaker.transformer.Transformer`. |
| | |
| | Args: |
| | name (str): The name of the `TransformStep`. |
| | step_args (_JobStepArguments): The arguments for the `TransformStep` definition. |
| | transformer (Transformer): A `sagemaker.transformer.Transformer` instance. |
| | inputs (TransformInput): A `sagemaker.inputs.TransformInput` instance. |
| | cache_config (CacheConfig): A `sagemaker.workflow.steps.CacheConfig` instance. |
| | display_name (str): The display name of the `TransformStep`. |
| | description (str): The description of the `TransformStep`. |
| | depends_on (List[Union[str, Step, StepCollection]]): A list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that this `TransformStep` |
| | depends on. |
| | retry_policies (List[RetryPolicy]): A list of retry policies. |
| | """ |
| | super(TransformStep, self).__init__( |
| | name, StepTypeEnum.TRANSFORM, display_name, description, depends_on, retry_policies |
| | ) |
| |
|
| | if not (step_args is not None) ^ (transformer is not None): |
| | raise ValueError("either step_args or transformer need to be given, but not both.") |
| |
|
| | if step_args: |
| | from sagemaker.workflow.utilities import validate_step_args_input |
| |
|
| | validate_step_args_input( |
| | step_args=step_args, |
| | expected_caller={Session.transform.__name__}, |
| | error_message="The step_args of TransformStep must be obtained " |
| | "from transformer.transform().", |
| | ) |
| |
|
| | self.step_args = step_args.args if step_args else None |
| | self.transformer = transformer |
| | self.inputs = inputs |
| | self.cache_config = cache_config |
| | self._properties = Properties(step_name=name, shape_name="DescribeTransformJobResponse") |
| |
|
| | if not self.step_args: |
| | if inputs is None: |
| | raise ValueError("Inputs can't be None when transformer is given.") |
| | warnings.warn( |
| | ( |
| | 'We are deprecating the instantiation of TransformStep using "transformer".' |
| | 'Instead, simply using "step_args".' |
| | ), |
| | DeprecationWarning, |
| | ) |
| |
|
| | @property |
| | def arguments(self) -> RequestType: |
| | """The arguments dictionary that is used to call `create_transform_job`. |
| | |
| | NOTE: The `CreateTransformJob` request is not quite the args list that workflow needs. |
| | `TransformJobName` and `ExperimentConfig` cannot be included in the arguments. |
| | """ |
| | if self.step_args: |
| | request_dict = self.step_args |
| | else: |
| | transform_args = _TransformJob._get_transform_args( |
| | transformer=self.transformer, |
| | data=self.inputs.data, |
| | data_type=self.inputs.data_type, |
| | content_type=self.inputs.content_type, |
| | compression_type=self.inputs.compression_type, |
| | split_type=self.inputs.split_type, |
| | input_filter=self.inputs.input_filter, |
| | output_filter=self.inputs.output_filter, |
| | join_source=self.inputs.join_source, |
| | model_client_config=self.inputs.model_client_config, |
| | experiment_config=dict(), |
| | batch_data_capture_config=self.inputs.batch_data_capture_config, |
| | ) |
| | request_dict = self.transformer.sagemaker_session._get_transform_request( |
| | **transform_args |
| | ) |
| |
|
| | request_dict.pop("TransformJobName", None) |
| | Step._trim_experiment_config(request_dict) |
| |
|
| | return request_dict |
| |
|
| | @property |
| | def properties(self): |
| | """A `Properties` object representing the `DescribeTransformJobResponse` data model.""" |
| | return self._properties |
| |
|
| | def to_request(self) -> RequestType: |
| | """Updates the dictionary with cache configuration.""" |
| | request_dict = super().to_request() |
| | if self.cache_config: |
| | request_dict.update(self.cache_config.config) |
| |
|
| | return request_dict |
| |
|
| |
|
| | class ProcessingStep(ConfigurableRetryStep): |
| | """`ProcessingStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_args: _JobStepArguments = None, |
| | processor: Processor = None, |
| | display_name: str = None, |
| | description: str = None, |
| | inputs: List[ProcessingInput] = None, |
| | outputs: List[ProcessingOutput] = None, |
| | job_arguments: List[str] = None, |
| | code: str = None, |
| | property_files: List[PropertyFile] = None, |
| | cache_config: CacheConfig = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: List[RetryPolicy] = None, |
| | kms_key=None, |
| | ): |
| | """Construct a `ProcessingStep`, given a `Processor` instance. |
| | |
| | In addition to the `Processor` instance, the other arguments are those that are supplied to |
| | the `process` method of the `sagemaker.processing.Processor`. |
| | |
| | Args: |
| | name (str): The name of the `ProcessingStep`. |
| | step_args (_JobStepArguments): The arguments for the `ProcessingStep` definition. |
| | processor (Processor): A `sagemaker.processing.Processor` instance. |
| | display_name (str): The display name of the `ProcessingStep`. |
| | description (str): The description of the `ProcessingStep` |
| | inputs (List[ProcessingInput]): A list of `sagemaker.processing.ProcessorInput` |
| | instances. Defaults to `None`. |
| | outputs (List[ProcessingOutput]): A list of `sagemaker.processing.ProcessorOutput` |
| | instances. Defaults to `None`. |
| | job_arguments (List[str]): A list of strings to be passed into the processing job. |
| | Defaults to `None`. |
| | code (str): This can be an S3 URI or a local path to a file with the framework |
| | script to run. Defaults to `None`. |
| | property_files (List[PropertyFile]): A list of property files that workflow looks |
| | for and resolves from the configured processing output list. |
| | cache_config (CacheConfig): A `sagemaker.workflow.steps.CacheConfig` instance. |
| | depends_on (List[Union[str, Step, StepCollection]]): A list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that this `ProcessingStep` |
| | depends on. |
| | retry_policies (List[RetryPolicy]): A list of retry policies. |
| | kms_key (str): The ARN of the KMS key that is used to encrypt the |
| | user code file. Defaults to `None`. |
| | """ |
| | super(ProcessingStep, self).__init__( |
| | name, StepTypeEnum.PROCESSING, display_name, description, depends_on, retry_policies |
| | ) |
| |
|
| | if not (step_args is not None) ^ (processor is not None): |
| | raise ValueError("either step_args or processor need to be given, but not both.") |
| |
|
| | if step_args: |
| | from sagemaker.workflow.utilities import validate_step_args_input |
| |
|
| | validate_step_args_input( |
| | step_args=step_args, |
| | expected_caller={Session.process.__name__}, |
| | error_message="The step_args of ProcessingStep must be obtained from processor.run().", |
| | ) |
| |
|
| | self.step_args = step_args.args if step_args else None |
| | self.processor = processor |
| | self.inputs = inputs |
| | self.outputs = outputs |
| | self.job_arguments = job_arguments |
| | self.code = code |
| | self.property_files = property_files or [] |
| | self.job_name = None |
| | self.kms_key = kms_key |
| | self.cache_config = cache_config |
| | self._properties = Properties(step_name=name, shape_name="DescribeProcessingJobResponse") |
| |
|
| | if not self.step_args: |
| | |
| | |
| | |
| | self.processor.arguments = job_arguments |
| |
|
| | if code: |
| | if is_pipeline_variable(code): |
| | raise ValueError( |
| | "code argument has to be a valid S3 URI or local file path " |
| | + "rather than a pipeline variable" |
| | ) |
| | code_url = urlparse(code) |
| | if code_url.scheme == "" or code_url.scheme == "file": |
| | |
| | |
| | |
| | |
| | |
| | |
| | self.job_name = self._generate_code_upload_path() |
| |
|
| | warnings.warn( |
| | ( |
| | 'We are deprecating the instantiation of ProcessingStep using "processor".' |
| | 'Instead, simply using "step_args".' |
| | ), |
| | DeprecationWarning, |
| | ) |
| |
|
| | @property |
| | def arguments(self) -> RequestType: |
| | """The arguments dictionary that is used to call `create_processing_job`. |
| | |
| | NOTE: The `CreateProcessingJob` request is not quite the args list that workflow needs. |
| | `ProcessingJobName` and `ExperimentConfig` cannot be included in the arguments. |
| | """ |
| | if self.step_args: |
| | request_dict = self.step_args |
| | else: |
| | normalized_inputs, normalized_outputs = self.processor._normalize_args( |
| | job_name=self.job_name, |
| | arguments=self.job_arguments, |
| | inputs=self.inputs, |
| | outputs=self.outputs, |
| | code=self.code, |
| | kms_key=self.kms_key, |
| | ) |
| | process_args = ProcessingJob._get_process_args( |
| | self.processor, normalized_inputs, normalized_outputs, experiment_config=dict() |
| | ) |
| | request_dict = self.processor.sagemaker_session._get_process_request(**process_args) |
| |
|
| | request_dict.pop("ProcessingJobName", None) |
| | Step._trim_experiment_config(request_dict) |
| |
|
| | return request_dict |
| |
|
| | @property |
| | def properties(self): |
| | """A `Properties` object representing the `DescribeProcessingJobResponse` data model.""" |
| | return self._properties |
| |
|
| | def to_request(self) -> RequestType: |
| | """Get the request structure for workflow service calls.""" |
| | request_dict = super(ProcessingStep, self).to_request() |
| | if self.cache_config: |
| | request_dict.update(self.cache_config.config) |
| | if self.property_files: |
| | request_dict["PropertyFiles"] = [ |
| | property_file.expr for property_file in self.property_files |
| | ] |
| | return request_dict |
| |
|
| | def _generate_code_upload_path(self) -> str: |
| | """Generate an upload path for local processing scripts based on its contents.""" |
| | from sagemaker.workflow.utilities import hash_file |
| |
|
| | code_hash = hash_file(self.code) |
| | return f"{self.name}-{code_hash}"[:1024] |
| |
|
| |
|
| | class TuningStep(ConfigurableRetryStep): |
| | """`TuningStep` for SageMaker Pipelines Workflows.""" |
| |
|
| | def __init__( |
| | self, |
| | name: str, |
| | step_args: _JobStepArguments = None, |
| | tuner: HyperparameterTuner = None, |
| | display_name: str = None, |
| | description: str = None, |
| | inputs=None, |
| | job_arguments: List[str] = None, |
| | cache_config: CacheConfig = None, |
| | depends_on: Optional[List[Union[str, Step, "StepCollection"]]] = None, |
| | retry_policies: List[RetryPolicy] = None, |
| | ): |
| | """Construct a `TuningStep`, given a `HyperparameterTuner` instance. |
| | |
| | In addition to the `HyperparameterTuner` instance, the other arguments are those |
| | that are supplied to the `fit` method of the `sagemaker.tuner.HyperparameterTuner`. |
| | |
| | Args: |
| | name (str): The name of the `TuningStep`. |
| | step_args (_JobStepArguments): The arguments for the `TuningStep` definition. |
| | tuner (HyperparameterTuner): A `sagemaker.tuner.HyperparameterTuner` instance. |
| | display_name (str): The display name of the `TuningStep`. |
| | description (str): The description of the `TuningStep`. |
| | inputs: Information about the training data. Please refer to the |
| | `fit()` method of the associated estimator, as this can take |
| | any of the following forms: |
| | |
| | * (str) - The S3 location where training data is saved. |
| | * (dict[str, str] or dict[str, sagemaker.inputs.TrainingInput]) - |
| | If using multiple channels for training data, you can specify |
| | a dictionary mapping channel names to strings or |
| | :func:`~sagemaker.inputs.TrainingInput` objects. |
| | * (sagemaker.inputs.TrainingInput) - Channel configuration for S3 data sources |
| | that can provide additional information about the training dataset. |
| | See :func:`sagemaker.inputs.TrainingInput` for full details. |
| | * (sagemaker.session.FileSystemInput) - channel configuration for |
| | a file system data source that can provide additional information as well as |
| | the path to the training dataset. |
| | * (sagemaker.amazon.amazon_estimator.RecordSet) - A collection of |
| | Amazon :class:~`Record` objects serialized and stored in S3. |
| | For use with an estimator for an Amazon algorithm. |
| | * (sagemaker.amazon.amazon_estimator.FileSystemRecordSet) - |
| | Amazon SageMaker channel configuration for a file system data source for |
| | Amazon algorithms. |
| | * (list[sagemaker.amazon.amazon_estimator.RecordSet]) - A list of |
| | :class:~`sagemaker.amazon.amazon_estimator.RecordSet` objects, |
| | where each instance is a different channel of training data. |
| | * (list[sagemaker.amazon.amazon_estimator.FileSystemRecordSet]) - A list of |
| | :class:~`sagemaker.amazon.amazon_estimator.FileSystemRecordSet` objects, |
| | where each instance is a different channel of training data. |
| | job_arguments (List[str]): A list of strings to be passed into the processing job. |
| | Defaults to `None`. |
| | cache_config (CacheConfig): A `sagemaker.workflow.steps.CacheConfig` instance. |
| | depends_on (List[Union[str, Step, StepCollection]]): A list of `Step`/`StepCollection` |
| | names or `Step` instances or `StepCollection` instances that this `TuningStep` |
| | depends on. |
| | retry_policies (List[RetryPolicy]): A list of retry policies. |
| | """ |
| | super(TuningStep, self).__init__( |
| | name, StepTypeEnum.TUNING, display_name, description, depends_on, retry_policies |
| | ) |
| |
|
| | if not (step_args is not None) ^ (tuner is not None): |
| | raise ValueError("either step_args or tuner need to be given, but not both.") |
| |
|
| | if step_args: |
| | from sagemaker.workflow.utilities import validate_step_args_input |
| |
|
| | validate_step_args_input( |
| | step_args=step_args, |
| | expected_caller={Session.create_tuning_job.__name__}, |
| | error_message="The step_args of TuningStep must be obtained from tuner.fit().", |
| | ) |
| |
|
| | self.step_args = step_args.args if step_args else None |
| | self.tuner = tuner |
| | self.inputs = inputs |
| | self.job_arguments = job_arguments |
| | self._properties = Properties( |
| | step_name=name, |
| | shape_names=[ |
| | "DescribeHyperParameterTuningJobResponse", |
| | "ListTrainingJobsForHyperParameterTuningJobResponse", |
| | ], |
| | ) |
| | self.cache_config = cache_config |
| |
|
| | if not self.step_args: |
| | warnings.warn( |
| | ( |
| | 'We are deprecating the instantiation of TuningStep using "tuner".' |
| | 'Instead, simply using "step_args".' |
| | ), |
| | DeprecationWarning, |
| | ) |
| |
|
| | @property |
| | def arguments(self) -> RequestType: |
| | """The arguments dictionary that is used to call `create_hyper_parameter_tuning_job`. |
| | |
| | NOTE: The `CreateHyperParameterTuningJob` request is not quite the |
| | args list that workflow needs. |
| | The `HyperParameterTuningJobName` attribute cannot be included. |
| | """ |
| | if self.step_args: |
| | request_dict = self.step_args |
| | else: |
| | if self.tuner.estimator is not None: |
| | self.tuner.estimator._prepare_for_training() |
| | else: |
| | for _, estimator in self.tuner.estimator_dict.items(): |
| | estimator._prepare_for_training() |
| |
|
| | self.tuner._prepare_for_tuning() |
| | tuner_args = _TuningJob._get_tuner_args(self.tuner, self.inputs) |
| | request_dict = self.tuner.sagemaker_session._get_tuning_request(**tuner_args) |
| |
|
| | request_dict.pop("HyperParameterTuningJobName", None) |
| | return request_dict |
| |
|
| | @property |
| | def properties(self): |
| | """A `Properties` object |
| | |
| | A `Properties` object representing `DescribeHyperParameterTuningJobResponse` and |
| | `ListTrainingJobsForHyperParameterTuningJobResponse` data model. |
| | """ |
| | return self._properties |
| |
|
| | def to_request(self) -> RequestType: |
| | """Updates the dictionary with cache configuration.""" |
| | request_dict = super().to_request() |
| | if self.cache_config: |
| | request_dict.update(self.cache_config.config) |
| |
|
| | return request_dict |
| |
|
| | def get_top_model_s3_uri(self, top_k: int, s3_bucket: str, prefix: str = "") -> Join: |
| | """Get the model artifact S3 URI from the top performing training jobs. |
| | |
| | Args: |
| | top_k (int): The index of the top performing training job |
| | tuning step stores up to 50 top performing training jobs. |
| | A valid top_k value is from 0 to 49. The best training job |
| | model is at index 0. |
| | s3_bucket (str): The S3 bucket to store the training job output artifact. |
| | prefix (str): The S3 key prefix to store the training job output artifact. |
| | """ |
| | values = ["s3:/", s3_bucket] |
| | if prefix != "" and prefix is not None: |
| | values.append(prefix) |
| |
|
| | return Join( |
| | on="/", |
| | values=values |
| | + [ |
| | self.properties.TrainingJobSummaries[top_k].TrainingJobName, |
| | "output/model.tar.gz", |
| | ], |
| | ) |
| |
|