| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """Local Pipeline Executor""" |
| | from __future__ import absolute_import |
| | from abc import ABC, abstractmethod |
| |
|
| | import json |
| | from copy import deepcopy |
| | from datetime import datetime |
| | from typing import Dict, List, Union |
| | from botocore.exceptions import ClientError |
| |
|
| | from sagemaker.workflow.conditions import ConditionTypeEnum |
| | from sagemaker.workflow.steps import StepTypeEnum, Step |
| | from sagemaker.workflow.step_collections import StepCollection |
| | from sagemaker.workflow.entities import PipelineVariable |
| | from sagemaker.workflow.parameters import Parameter |
| | from sagemaker.workflow.functions import Join, JsonGet, PropertyFile |
| | from sagemaker.workflow.properties import Properties |
| | from sagemaker.workflow.execution_variables import ExecutionVariable, ExecutionVariables |
| | from sagemaker.workflow.pipeline import PipelineGraph |
| | from sagemaker.local.exceptions import StepExecutionException |
| | from sagemaker.local.utils import get_using_dot_notation |
| | from sagemaker.utils import unique_name_from_base |
| | from sagemaker.s3 import parse_s3_url, s3_path_join |
| |
|
| |
|
| | PRIMITIVES = (str, int, bool, float) |
| | BINARY_CONDITION_TYPES = ( |
| | ConditionTypeEnum.EQ.value, |
| | ConditionTypeEnum.GT.value, |
| | ConditionTypeEnum.GTE.value, |
| | ConditionTypeEnum.LT.value, |
| | ConditionTypeEnum.LTE.value, |
| | ) |
| |
|
| |
|
| | class LocalPipelineExecutor(object): |
| | """An executor that executes SageMaker Pipelines locally.""" |
| |
|
| | def __init__(self, execution, sagemaker_session): |
| | """Initialize StepExecutor. |
| | |
| | Args: |
| | sagemaker_session (sagemaker.session.Session): a session to use to read configurations |
| | from, and use its boto client. |
| | """ |
| | self.sagemaker_session = sagemaker_session |
| | self.execution = execution |
| | self.pipeline_dag = PipelineGraph.from_pipeline(self.execution.pipeline) |
| | self.local_sagemaker_client = self.sagemaker_session.sagemaker_client |
| | self._blocked_steps = set() |
| | self._step_executor_factory = _StepExecutorFactory(self) |
| |
|
| | def execute(self): |
| | """Execute a local pipeline.""" |
| | try: |
| | for step in self.pipeline_dag: |
| | if step.name not in self._blocked_steps: |
| | self._execute_step(step) |
| | except StepExecutionException as e: |
| | self.execution.update_execution_failure(e.step_name, e.message) |
| | else: |
| | self.execution.update_execution_success() |
| | return self.execution |
| |
|
| | def _execute_step(self, step): |
| | """Execute a local pipeline step.""" |
| | self.execution.mark_step_executing(step.name) |
| | step_properties = self._step_executor_factory.get(step).execute() |
| | self.execution.update_step_properties(step.name, step_properties) |
| |
|
| | def evaluate_step_arguments(self, step): |
| | """Parses and evaluate step arguments.""" |
| | return self._parse_arguments(step.arguments, step.name) |
| |
|
| | def _parse_arguments(self, obj, step_name): |
| | """Parse and evaluate arguments field""" |
| | if isinstance(obj, dict): |
| | obj_copy = deepcopy(obj) |
| | for k, v in obj.items(): |
| | obj_copy[k] = self._parse_arguments(v, step_name) |
| | return obj_copy |
| | if isinstance(obj, list): |
| | list_copy = [] |
| | for item in obj: |
| | list_copy.append(self._parse_arguments(item, step_name)) |
| | return list_copy |
| | if isinstance(obj, PipelineVariable): |
| | return self.evaluate_pipeline_variable(obj, step_name) |
| | return obj |
| |
|
| | def evaluate_pipeline_variable(self, pipeline_variable, step_name): |
| | """Evaluate pipeline variable runtime value.""" |
| | value = None |
| | if isinstance(pipeline_variable, PRIMITIVES): |
| | value = pipeline_variable |
| | elif isinstance(pipeline_variable, Parameter): |
| | value = self.execution.pipeline_parameters.get(pipeline_variable.name) |
| | elif isinstance(pipeline_variable, Join): |
| | evaluated = [ |
| | str(self.evaluate_pipeline_variable(v, step_name)) for v in pipeline_variable.values |
| | ] |
| | value = pipeline_variable.on.join(evaluated) |
| | elif isinstance(pipeline_variable, Properties): |
| | value = self._evaluate_property_reference(pipeline_variable, step_name) |
| | elif isinstance(pipeline_variable, ExecutionVariable): |
| | value = self._evaluate_execution_variable(pipeline_variable) |
| | elif isinstance(pipeline_variable, JsonGet): |
| | value = self._evaluate_json_get_function(pipeline_variable, step_name) |
| | else: |
| | self.execution.update_step_failure( |
| | step_name, f"Unrecognized pipeline variable {pipeline_variable.expr}." |
| | ) |
| |
|
| | if value is None: |
| | self.execution.update_step_failure(step_name, f"{pipeline_variable.expr} is undefined.") |
| | return value |
| |
|
| | def _evaluate_property_reference(self, pipeline_variable, step_name): |
| | """Evaluate property reference runtime value.""" |
| | try: |
| | referenced_step_name = pipeline_variable.step_name |
| | step_properties = self.execution.step_execution.get(referenced_step_name).properties |
| | return get_using_dot_notation(step_properties, pipeline_variable.path) |
| | except ValueError: |
| | self.execution.update_step_failure(step_name, f"{pipeline_variable.expr} is undefined.") |
| |
|
| | def _evaluate_execution_variable(self, pipeline_variable): |
| | """Evaluate pipeline execution variable runtime value.""" |
| | if pipeline_variable in ( |
| | ExecutionVariables.PIPELINE_NAME, |
| | ExecutionVariables.PIPELINE_ARN, |
| | ): |
| | return self.execution.pipeline.name |
| | if pipeline_variable in ( |
| | ExecutionVariables.PIPELINE_EXECUTION_ID, |
| | ExecutionVariables.PIPELINE_EXECUTION_ARN, |
| | ): |
| | return self.execution.pipeline_execution_name |
| | if pipeline_variable == ExecutionVariables.START_DATETIME: |
| | return self.execution.creation_time |
| | if pipeline_variable == ExecutionVariables.CURRENT_DATETIME: |
| | return datetime.now() |
| | return None |
| |
|
| | def _evaluate_json_get_function(self, pipeline_variable, step_name): |
| | """Evaluate join function runtime value.""" |
| | property_file_reference = pipeline_variable.property_file |
| | property_file = None |
| | if isinstance(property_file_reference, str): |
| | processing_step = self.pipeline_dag.step_map[pipeline_variable.step_name] |
| | for file in processing_step.property_files: |
| | if file.name == property_file_reference: |
| | property_file = file |
| | break |
| | elif isinstance(property_file_reference, PropertyFile): |
| | property_file = property_file_reference |
| | processing_step_response = self.execution.step_execution.get( |
| | pipeline_variable.step_name |
| | ).properties |
| | if ( |
| | "ProcessingOutputConfig" not in processing_step_response |
| | or "Outputs" not in processing_step_response["ProcessingOutputConfig"] |
| | ): |
| | self.execution.update_step_failure( |
| | step_name, |
| | f"Step '{pipeline_variable.step_name}' does not yet contain processing outputs.", |
| | ) |
| | processing_output_s3_bucket = processing_step_response["ProcessingOutputConfig"]["Outputs"][ |
| | property_file.output_name |
| | ]["S3Output"]["S3Uri"] |
| | try: |
| | s3_bucket, s3_key_prefix = parse_s3_url(processing_output_s3_bucket) |
| | file_content = self.sagemaker_session.read_s3_file( |
| | s3_bucket, s3_path_join(s3_key_prefix, property_file.path) |
| | ) |
| | file_json = json.loads(file_content) |
| | return get_using_dot_notation(file_json, pipeline_variable.json_path) |
| | except ClientError as e: |
| | self.execution.update_step_failure( |
| | step_name, |
| | f"Received an error while file reading file '{property_file.path}' from S3: " |
| | f"{e.response.get('Code')}: {e.response.get('Message')}", |
| | ) |
| | except json.JSONDecodeError: |
| | self.execution.update_step_failure( |
| | step_name, |
| | f"Contents of property file '{property_file.name}' are not in valid JSON format.", |
| | ) |
| | except ValueError: |
| | self.execution.update_step_failure( |
| | step_name, f"Invalid json path '{pipeline_variable.json_path}'" |
| | ) |
| |
|
| |
|
| | class _StepExecutor(ABC): |
| | """An abstract base class for step executors running steps locally""" |
| |
|
| | def __init__(self, pipeline_executor: LocalPipelineExecutor, step: Step): |
| | self.pipline_executor = pipeline_executor |
| | self.step = step |
| |
|
| | @abstractmethod |
| | def execute(self) -> Dict: |
| | """Execute a pipeline step locally |
| | |
| | Returns: |
| | A dictionary as properties of the current step |
| | """ |
| |
|
| | def _convert_list_to_dict(self, dictionary: dict, path_to_list: str, reducing_key: str): |
| | """Convert list into dictionary using a field inside list elements as the keys. |
| | |
| | Raises RuntimeError if given list not able to be converted into a map based on given key. |
| | """ |
| |
|
| | try: |
| | list_to_convert = get_using_dot_notation(dictionary, path_to_list) |
| | except ValueError: |
| | raise RuntimeError(f"{path_to_list} does not exist in {dictionary}") |
| | if not isinstance(list_to_convert, list): |
| | raise RuntimeError( |
| | f"Element at path {path_to_list} is not a list. Actual type {type(list_to_convert)}" |
| | ) |
| | converted_map = {} |
| | for element in list_to_convert: |
| | if not isinstance(element, dict): |
| | raise RuntimeError( |
| | f"Cannot convert element of type {type(element)} into dictionary entry" |
| | ) |
| | converted_map[element[reducing_key]] = element |
| | return converted_map |
| |
|
| |
|
| | class _TrainingStepExecutor(_StepExecutor): |
| | """Executor class to execute TrainingStep locally""" |
| |
|
| | def execute(self): |
| | job_name = unique_name_from_base(self.step.name) |
| | step_arguments = self.pipline_executor.evaluate_step_arguments(self.step) |
| | try: |
| | self.pipline_executor.local_sagemaker_client.create_training_job( |
| | job_name, **step_arguments |
| | ) |
| | return self.pipline_executor.local_sagemaker_client.describe_training_job(job_name) |
| | except Exception as e: |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, f"{type(e).__name__}: {str(e)}" |
| | ) |
| |
|
| |
|
| | class _ProcessingStepExecutor(_StepExecutor): |
| | """Executor class to execute ProcessingStep locally""" |
| |
|
| | def execute(self): |
| | job_name = unique_name_from_base(self.step.name) |
| | step_arguments = self.pipline_executor.evaluate_step_arguments(self.step) |
| | try: |
| | self.pipline_executor.local_sagemaker_client.create_processing_job( |
| | job_name, **step_arguments |
| | ) |
| | job_describe_response = ( |
| | self.pipline_executor.local_sagemaker_client.describe_processing_job(job_name) |
| | ) |
| | if ( |
| | "ProcessingOutputConfig" in job_describe_response |
| | and "Outputs" in job_describe_response["ProcessingOutputConfig"] |
| | ): |
| | job_describe_response["ProcessingOutputConfig"][ |
| | "Outputs" |
| | ] = self._convert_list_to_dict( |
| | job_describe_response, "ProcessingOutputConfig.Outputs", "OutputName" |
| | ) |
| | if "ProcessingInputs" in job_describe_response: |
| | job_describe_response["ProcessingInputs"] = self._convert_list_to_dict( |
| | job_describe_response, "ProcessingInputs", "InputName" |
| | ) |
| | return job_describe_response |
| |
|
| | except Exception as e: |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, f"{type(e).__name__}: {str(e)}" |
| | ) |
| |
|
| |
|
| | class _ConditionStepExecutor(_StepExecutor): |
| | """Executor class to execute ConditionStep locally""" |
| |
|
| | def execute(self): |
| | def _block_all_downstream_steps(steps: List[Union[Step, StepCollection]]): |
| | steps_to_block = set() |
| | for step in steps: |
| | steps_to_block.update(self.pipline_executor.pipeline_dag.get_steps_in_sub_dag(step)) |
| | self.pipline_executor._blocked_steps.update(steps_to_block) |
| |
|
| | if_steps = self.step.if_steps |
| | else_steps = self.step.else_steps |
| | step_only_arguments = self.pipline_executor._parse_arguments( |
| | self.step.step_only_arguments, self.step.name |
| | ) |
| |
|
| | outcome = self._evaluate_conjunction(step_only_arguments["Conditions"]) |
| |
|
| | if not outcome: |
| | _block_all_downstream_steps(if_steps) |
| | else: |
| | _block_all_downstream_steps(else_steps) |
| |
|
| | return dict(Outcome=outcome) |
| |
|
| | def _evaluate_conjunction(self, conditions: List[Dict]) -> bool: |
| | """Evaluate conditions of current conditionStep. |
| | |
| | Args: |
| | List of dictionaries representing conditions as request |
| | |
| | Returns: |
| | True if the conjunction expression is true, |
| | False otherwise. |
| | """ |
| | for condition in conditions: |
| | if not self._resolve_condition(condition): |
| | return False |
| | return True |
| |
|
| | def _resolve_condition(self, condition: dict) -> bool: |
| | """Resolve given condition. |
| | |
| | Args: |
| | Dictionary representing given condition as request |
| | |
| | Returns: |
| | True if given condition evaluated as true, |
| | False otherwise. |
| | """ |
| |
|
| | condition_type = condition["Type"] |
| | outcome = None |
| | if condition_type in BINARY_CONDITION_TYPES: |
| | outcome = self._resolve_binary_condition(condition, condition_type) |
| | elif condition_type == ConditionTypeEnum.NOT.value: |
| | outcome = self._resolve_not_condition(condition) |
| | elif condition_type == ConditionTypeEnum.OR.value: |
| | outcome = self._resolve_or_condition(condition) |
| | elif condition_type == ConditionTypeEnum.IN.value: |
| | outcome = self._resolve_in_condition(condition) |
| | else: |
| | raise NotImplementedError(f"Condition of type [{condition_type}] is not supported.") |
| |
|
| | return outcome |
| |
|
| | def _resolve_binary_condition(self, binary_condition: dict, binary_condition_type: str): |
| | """Resolve given binary condition. |
| | |
| | Args: |
| | Dictionary representing given binary condition as request |
| | |
| | Returns: |
| | True if given binary condition evaluated as true, |
| | False otherwise. |
| | """ |
| |
|
| | left_value = binary_condition["LeftValue"] |
| | right_value = binary_condition["RightValue"] |
| | try: |
| | outcome = None |
| | if binary_condition_type == ConditionTypeEnum.EQ.value: |
| | if not isinstance(left_value, type(right_value)) and not isinstance( |
| | right_value, type(left_value) |
| | ): |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, |
| | f"LeftValue [{left_value}] of type [{type(left_value)}] and " |
| | + f"RightValue [{right_value}] of type [{type(right_value)}] " |
| | + "are not of the same type.", |
| | ) |
| | outcome = left_value == right_value |
| | elif binary_condition_type == ConditionTypeEnum.GT.value: |
| | outcome = left_value > right_value |
| | elif binary_condition_type == ConditionTypeEnum.GTE.value: |
| | outcome = left_value >= right_value |
| | elif binary_condition_type == ConditionTypeEnum.LT.value: |
| | outcome = left_value < right_value |
| | elif binary_condition_type == ConditionTypeEnum.LTE.value: |
| | outcome = left_value <= right_value |
| | else: |
| | raise NotImplementedError( |
| | f"Binary condition of type [{binary_condition_type}] is not supported" |
| | ) |
| | return outcome |
| |
|
| | except TypeError: |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, |
| | f"Condition of type [{binary_condition_type}] not supported between " |
| | + f"[{left_value}] of type [{type(left_value)}] and [{right_value}] " |
| | + f"of type [{type(right_value)}]", |
| | ) |
| |
|
| | def _resolve_not_condition(self, not_condition: dict): |
| | """Resolve given ConditionNot. |
| | |
| | Args: |
| | Dictionary representing given ConditionNot as request |
| | |
| | Returns: |
| | True if given ConditionNot evaluated as true, |
| | False otherwise. |
| | """ |
| | return not self._resolve_condition(not_condition["Expression"]) |
| |
|
| | def _resolve_or_condition(self, or_condition: dict): |
| | """Resolve given ConditionOr. |
| | |
| | Args: |
| | Dictionary representing given ConditionOr as request |
| | |
| | Returns: |
| | True if given ConditionOr evaluated as true, |
| | False otherwise. |
| | """ |
| |
|
| | for condition in or_condition["Conditions"]: |
| | if self._resolve_condition(condition): |
| | return True |
| | return False |
| |
|
| | def _resolve_in_condition(self, in_condition: dict): |
| | """Resolve given ConditionIn. |
| | |
| | Args: |
| | Dictionary representing given ConditionIn as request |
| | |
| | Returns: |
| | True if given ConditionIn evaluated as true, |
| | False otherwise. |
| | """ |
| |
|
| | query_value = in_condition["QueryValue"] |
| | values = in_condition["Values"] |
| | return query_value in values |
| |
|
| |
|
| | class _TransformStepExecutor(_StepExecutor): |
| | """Executor class to execute TransformStep locally""" |
| |
|
| | def execute(self): |
| | job_name = unique_name_from_base(self.step.name) |
| | step_arguments = self.pipline_executor.evaluate_step_arguments(self.step) |
| | try: |
| | self.pipline_executor.local_sagemaker_client.create_transform_job( |
| | job_name, **step_arguments |
| | ) |
| | return self.pipline_executor.local_sagemaker_client.describe_transform_job(job_name) |
| | except Exception as e: |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, f"{type(e).__name__}: {str(e)}" |
| | ) |
| |
|
| |
|
| | class _CreateModelStepExecutor(_StepExecutor): |
| | """Executor class to execute CreateModelStep locally""" |
| |
|
| | def execute(self): |
| | model_name = unique_name_from_base(self.step.name) |
| | step_arguments = self.pipline_executor.evaluate_step_arguments(self.step) |
| | try: |
| | self.pipline_executor.local_sagemaker_client.create_model(model_name, **step_arguments) |
| | return self.pipline_executor.local_sagemaker_client.describe_model(model_name) |
| | except Exception as e: |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, f"{type(e).__name__}: {str(e)}" |
| | ) |
| |
|
| |
|
| | class _FailStepExecutor(_StepExecutor): |
| | """Executor class to execute FailStep locally""" |
| |
|
| | def execute(self): |
| | step_arguments = self.pipline_executor.evaluate_step_arguments(self.step) |
| |
|
| | error_message = step_arguments.get("ErrorMessage") |
| | self.pipline_executor.execution.update_step_properties( |
| | self.step.name, {"ErrorMessage": error_message} |
| | ) |
| | self.pipline_executor.execution.update_step_failure( |
| | self.step.name, step_arguments.get("ErrorMessage") |
| | ) |
| |
|
| |
|
| | class _StepExecutorFactory: |
| | """Factory class to generate executors for given step based on their types""" |
| |
|
| | def __init__(self, pipeline_executor: LocalPipelineExecutor): |
| | self.pipeline_executor = pipeline_executor |
| |
|
| | def get(self, step: Step) -> _StepExecutor: |
| | """Return corresponding step executor for given step""" |
| |
|
| | step_type = step.step_type |
| | step_executor = None |
| | if step_type == StepTypeEnum.TRAINING: |
| | step_executor = _TrainingStepExecutor(self.pipeline_executor, step) |
| | elif step_type == StepTypeEnum.PROCESSING: |
| | step_executor = _ProcessingStepExecutor(self.pipeline_executor, step) |
| | elif step_type == StepTypeEnum.TRANSFORM: |
| | step_executor = _TransformStepExecutor(self.pipeline_executor, step) |
| | elif step_type == StepTypeEnum.CREATE_MODEL: |
| | step_executor = _CreateModelStepExecutor(self.pipeline_executor, step) |
| | elif step_type == StepTypeEnum.FAIL: |
| | step_executor = _FailStepExecutor(self.pipeline_executor, step) |
| | elif step_type == StepTypeEnum.CONDITION: |
| | step_executor = _ConditionStepExecutor(self.pipeline_executor, step) |
| | else: |
| | self.pipeline_executor.execution.update_step_failure( |
| | step.name, f"Unsupported step type {step_type} to execute." |
| | ) |
| | return step_executor |
| |
|