| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """Placeholder docstring""" |
| | from __future__ import absolute_import |
| |
|
| | import logging |
| | from typing import Union, Optional, Dict |
| |
|
| | from packaging.version import Version |
| |
|
| | from sagemaker.estimator import Framework, EstimatorBase |
| | from sagemaker.fw_utils import ( |
| | framework_name_from_image, |
| | framework_version_from_tag, |
| | python_deprecation_warning, |
| | validate_version_or_image_args, |
| | validate_distribution, |
| | ) |
| | from sagemaker.pytorch import defaults |
| | from sagemaker.pytorch.model import PyTorchModel |
| | from sagemaker.vpc_utils import VPC_CONFIG_DEFAULT |
| | from sagemaker.workflow.entities import PipelineVariable |
| |
|
| | logger = logging.getLogger("sagemaker") |
| |
|
| |
|
| | class PyTorch(Framework): |
| | """Handle end-to-end training and deployment of custom PyTorch code.""" |
| |
|
| | _framework_name = "pytorch" |
| | LAUNCH_PYTORCH_DDP_ENV_NAME = "sagemaker_pytorch_ddp_enabled" |
| | LAUNCH_TORCH_DISTRIBUTED_ENV_NAME = "sagemaker_torch_distributed_enabled" |
| | INSTANCE_TYPE_ENV_NAME = "sagemaker_instance_type" |
| |
|
| | def __init__( |
| | self, |
| | entry_point: Union[str, PipelineVariable], |
| | framework_version: Optional[str] = None, |
| | py_version: Optional[str] = None, |
| | source_dir: Optional[Union[str, PipelineVariable]] = None, |
| | hyperparameters: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| | image_uri: Optional[Union[str, PipelineVariable]] = None, |
| | distribution: Optional[Dict] = None, |
| | **kwargs |
| | ): |
| | """This ``Estimator`` executes a PyTorch script in a managed PyTorch execution environment. |
| | |
| | The managed PyTorch environment is an Amazon-built Docker container that executes functions |
| | defined in the supplied ``entry_point`` Python script within a SageMaker Training Job. |
| | |
| | Training is started by calling |
| | :meth:`~sagemaker.amazon.estimator.Framework.fit` on this Estimator. |
| | After training is complete, calling |
| | :meth:`~sagemaker.amazon.estimator.Framework.deploy` creates a hosted |
| | SageMaker endpoint and returns an |
| | :class:`~sagemaker.amazon.pytorch.model.PyTorchPredictor` instance that |
| | can be used to perform inference against the hosted model. |
| | |
| | Technical documentation on preparing PyTorch scripts for SageMaker |
| | training and using the PyTorch Estimator is available on the project |
| | home-page: https://github.com/aws/sagemaker-python-sdk |
| | |
| | Args: |
| | entry_point (str or PipelineVariable): Path (absolute or relative) to the |
| | Python source file which should be executed as the entry point to training. |
| | If ``source_dir`` is specified, then ``entry_point`` |
| | must point to a file located at the root of ``source_dir``. |
| | framework_version (str): PyTorch version you want to use for |
| | executing your model training code. Defaults to ``None``. Required unless |
| | ``image_uri`` is provided. List of supported versions: |
| | https://github.com/aws/deep-learning-containers/blob/master/available_images.md. |
| | py_version (str): Python version you want to use for executing your |
| | model training code. One of 'py2' or 'py3'. Defaults to ``None``. Required |
| | unless ``image_uri`` is provided. |
| | source_dir (str or PipelineVariable): Path (absolute, relative or an S3 URI) to |
| | a directory with any other training source code dependencies aside from the entry |
| | point file (default: None). If ``source_dir`` is an S3 URI, it must |
| | point to a tar.gz file. Structure within this directory are preserved |
| | when training on Amazon SageMaker. |
| | hyperparameters (dict[str, str] or dict[str, PipelineVariable]): Hyperparameters |
| | that will be used for training (default: None). The hyperparameters are made |
| | accessible as a dict[str, str] to the training code on |
| | SageMaker. For convenience, this accepts other types for keys |
| | and values, but ``str()`` will be called to convert them before |
| | training. |
| | image_uri (str or PipelineVariable): If specified, the estimator will use this image |
| | for training and hosting, instead of selecting the appropriate |
| | SageMaker official image based on framework_version and |
| | py_version. It can be an ECR url or dockerhub image and tag. |
| | Examples: |
| | * ``123412341234.dkr.ecr.us-west-2.amazonaws.com/my-custom-image:1.0`` |
| | * ``custom-image:latest`` |
| | |
| | If ``framework_version`` or ``py_version`` are ``None``, then |
| | ``image_uri`` is required. If also ``None``, then a ``ValueError`` |
| | will be raised. |
| | distribution (dict): A dictionary with information on how to run distributed training |
| | (default: None). Currently, the following are supported: |
| | distributed training with parameter servers, SageMaker Distributed (SMD) Data |
| | and Model Parallelism, and MPI. SMD Model Parallelism can only be used with MPI. |
| | |
| | **To enable the SageMaker distributed data parallelism:** |
| | |
| | .. code:: python |
| | |
| | { "smdistributed": { "dataparallel": { "enabled": True } } } |
| | |
| | .. seealso:: |
| | |
| | To learn more, see :ref:`sdp_api_docs_toc`. |
| | |
| | **To enable the SageMaker distributed model parallelism:** |
| | |
| | .. code:: python |
| | |
| | { |
| | "smdistributed": { |
| | "modelparallel": { |
| | "enabled":True, |
| | "parameters": { |
| | "partitions": 2, |
| | "microbatches": 4, |
| | "placement_strategy": "spread", |
| | "pipeline": "interleaved", |
| | "optimize": "speed", |
| | "ddp": True, |
| | } |
| | }, |
| | "mpi": { |
| | "enabled" : True, |
| | "processes_per_host" : 8, |
| | } |
| | } |
| | |
| | .. note:: |
| | |
| | The SageMaker distributed model parallel library internally uses MPI. |
| | In order to use model parallelism, MPI also must be enabled. |
| | |
| | .. seealso:: |
| | |
| | To learn more, see :ref:`smp_api_docs_toc`. |
| | |
| | .. seealso:: |
| | |
| | To find a complete list of parameters for SageMaker model parallelism, |
| | see :ref:`sm-sdk-modelparallel-general`. |
| | |
| | **To enable PyTorch DDP:** |
| | |
| | .. code:: python |
| | |
| | { |
| | "pytorchddp": { |
| | "enabled": True |
| | } |
| | } |
| | |
| | To learn more, see `Distributed PyTorch Training |
| | <https://sagemaker.readthedocs.io/en/stable/frameworks/pytorch/using_pytorch.html#distributed-pytorch-training>`_. |
| | |
| | **To enable Torch Distributed (for Trainium instances only):** |
| | |
| | .. code:: python |
| | |
| | { |
| | "torch_distributed": { |
| | "enabled": True |
| | } |
| | } |
| | |
| | To learn more, see `Distributed PyTorch Training on Trainium |
| | <https://sagemaker.readthedocs.io/en/stable/frameworks/pytorch/using_pytorch.html#distributed-pytorch-training-on-trainium>`_. |
| | |
| | **To enable MPI:** |
| | |
| | .. code:: python |
| | |
| | { |
| | "mpi": { |
| | "enabled": True |
| | } |
| | } |
| | |
| | To learn more, see `Training with Horovod |
| | <https://sagemaker.readthedocs.io/en/stable/frameworks/tensorflow/using_tf.html#training-with-horovod>`_. |
| | |
| | **To enable parameter server:** |
| | |
| | .. code:: python |
| | |
| | { |
| | "parameter_server": { |
| | "enabled": True |
| | } |
| | } |
| | |
| | To learn more, see `Training with parameter servers |
| | <https://sagemaker.readthedocs.io/en/stable/frameworks/tensorflow/using_tf.html#training-with-parameter-servers>`_. |
| | |
| | **kwargs: Additional kwargs passed to the :class:`~sagemaker.estimator.Framework` |
| | constructor. |
| | |
| | .. tip:: |
| | |
| | You can find additional parameters for initializing this class at |
| | :class:`~sagemaker.estimator.Framework` and |
| | :class:`~sagemaker.estimator.EstimatorBase`. |
| | """ |
| | validate_version_or_image_args(framework_version, py_version, image_uri) |
| | if py_version == "py2": |
| | logger.warning( |
| | python_deprecation_warning(self._framework_name, defaults.LATEST_PY2_VERSION) |
| | ) |
| | self.framework_version = framework_version |
| | self.py_version = py_version |
| |
|
| | if "enable_sagemaker_metrics" not in kwargs: |
| | |
| | if self.framework_version and Version(self.framework_version) >= Version("1.3"): |
| | kwargs["enable_sagemaker_metrics"] = True |
| |
|
| | super(PyTorch, self).__init__( |
| | entry_point, source_dir, hyperparameters, image_uri=image_uri, **kwargs |
| | ) |
| |
|
| | if "entry_point" not in kwargs: |
| | kwargs["entry_point"] = entry_point |
| |
|
| | if distribution is not None: |
| | distribution = validate_distribution( |
| | distribution, |
| | self.instance_groups, |
| | self._framework_name, |
| | framework_version, |
| | py_version, |
| | image_uri, |
| | kwargs, |
| | ) |
| |
|
| | self.distribution = distribution or {} |
| |
|
| | def _pytorch_distribution_configuration(self, distribution): |
| | """Returns a dict of distribution config for PyTorch training |
| | |
| | Args: |
| | distribution (dict): A dictionary with information on how to run distributed training. |
| | Returns: |
| | dict containing Pytorch DDP config |
| | """ |
| | distribution_config = {} |
| | pytorch_ddp_enabled = False |
| | torch_distributed_enabled = False |
| |
|
| | if "pytorchddp" in distribution: |
| | pytorch_ddp_enabled = distribution.get("pytorchddp").get("enabled", False) |
| | elif "torch_distributed" in distribution: |
| | torch_distributed_enabled = distribution.get("torch_distributed").get("enabled", False) |
| |
|
| | if pytorch_ddp_enabled: |
| | distribution_config[self.LAUNCH_PYTORCH_DDP_ENV_NAME] = pytorch_ddp_enabled |
| | if self.instance_type is not None: |
| | distribution_config[self.INSTANCE_TYPE_ENV_NAME] = self.instance_type |
| | elif torch_distributed_enabled: |
| | distribution_config[self.LAUNCH_TORCH_DISTRIBUTED_ENV_NAME] = torch_distributed_enabled |
| | if self.instance_type is not None: |
| | distribution_config[self.INSTANCE_TYPE_ENV_NAME] = self.instance_type |
| | else: |
| | distribution_config = self._distribution_configuration(distribution=distribution) |
| |
|
| | return distribution_config |
| |
|
| | def hyperparameters(self): |
| | """Return hyperparameters used by your custom PyTorch code during model training.""" |
| | hyperparameters = super(PyTorch, self).hyperparameters() |
| | additional_hyperparameters = self._pytorch_distribution_configuration( |
| | distribution=self.distribution |
| | ) |
| | hyperparameters.update( |
| | EstimatorBase._json_encode_hyperparameters(additional_hyperparameters) |
| | ) |
| | return hyperparameters |
| |
|
| | def create_model( |
| | self, |
| | model_server_workers=None, |
| | role=None, |
| | vpc_config_override=VPC_CONFIG_DEFAULT, |
| | entry_point=None, |
| | source_dir=None, |
| | dependencies=None, |
| | **kwargs |
| | ): |
| | """Create a SageMaker ``PyTorchModel`` object that can be deployed to an ``Endpoint``. |
| | |
| | Args: |
| | model_server_workers (int): Optional. The number of worker processes |
| | used by the inference server. If None, server will use one |
| | worker per vCPU. |
| | role (str): The ``ExecutionRoleArn`` IAM Role ARN for the ``Model``, |
| | which is also used during transform jobs. If not specified, the |
| | role from the Estimator will be used. |
| | vpc_config_override (dict[str, list[str]]): Optional override for VpcConfig set on |
| | the model. Default: use subnets and security groups from this Estimator. |
| | * 'Subnets' (list[str]): List of subnet ids. |
| | * 'SecurityGroupIds' (list[str]): List of security group ids. |
| | entry_point (str): Path (absolute or relative) to the local Python source file which |
| | should be executed as the entry point to training. If ``source_dir`` is specified, |
| | then ``entry_point`` must point to a file located at the root of ``source_dir``. |
| | If not specified, the training entry point is used. |
| | source_dir (str): Path (absolute or relative) to a directory with any other serving |
| | source code dependencies aside from the entry point file. |
| | If not specified, the model source directory from training is used. |
| | dependencies (list[str]): A list of paths to directories (absolute or relative) with |
| | any additional libraries that will be exported to the container. |
| | If not specified, the dependencies from training are used. |
| | This is not supported with "local code" in Local Mode. |
| | **kwargs: Additional kwargs passed to the :class:`~sagemaker.pytorch.model.PyTorchModel` |
| | constructor. |
| | |
| | Returns: |
| | sagemaker.pytorch.model.PyTorchModel: A SageMaker ``PyTorchModel`` |
| | object. See :func:`~sagemaker.pytorch.model.PyTorchModel` for full details. |
| | """ |
| | if "image_uri" not in kwargs: |
| | kwargs["image_uri"] = self.image_uri |
| |
|
| | kwargs["name"] = self._get_or_create_name(kwargs.get("name")) |
| |
|
| | return PyTorchModel( |
| | self.model_data, |
| | role or self.role, |
| | entry_point or self._model_entry_point(), |
| | framework_version=self.framework_version, |
| | py_version=self.py_version, |
| | source_dir=(source_dir or self._model_source_dir()), |
| | container_log_level=self.container_log_level, |
| | code_location=self.code_location, |
| | model_server_workers=model_server_workers, |
| | sagemaker_session=self.sagemaker_session, |
| | vpc_config=self.get_vpc_config(vpc_config_override), |
| | dependencies=(dependencies or self.dependencies), |
| | **kwargs |
| | ) |
| |
|
| | @classmethod |
| | def _prepare_init_params_from_job_description(cls, job_details, model_channel_name=None): |
| | """Convert the job description to init params that can be handled by the class constructor. |
| | |
| | Args: |
| | job_details: the returned job details from a describe_training_job |
| | API call. |
| | model_channel_name (str): Name of the channel where pre-trained |
| | model data will be downloaded. |
| | |
| | Returns: |
| | dictionary: The transformed init_params |
| | """ |
| | init_params = super(PyTorch, cls)._prepare_init_params_from_job_description( |
| | job_details, model_channel_name |
| | ) |
| | image_uri = init_params.pop("image_uri") |
| | framework, py_version, tag, _ = framework_name_from_image(image_uri) |
| |
|
| | if tag is None: |
| | framework_version = None |
| | else: |
| | framework_version = framework_version_from_tag(tag) |
| | init_params["framework_version"] = framework_version |
| | init_params["py_version"] = py_version |
| |
|
| | if not framework: |
| | |
| | |
| | init_params["image_uri"] = image_uri |
| | return init_params |
| |
|
| | if framework != cls._framework_name: |
| | raise ValueError( |
| | "Training job: {} didn't use image for requested framework".format( |
| | job_details["TrainingJobName"] |
| | ) |
| | ) |
| |
|
| | return init_params |
| |
|