# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"). You # may not use this file except in compliance with the License. A copy of # the License is located at # # http://aws.amazon.com/apache2.0/ # # or in the "license" file accompanying this file. This file is # distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF # ANY KIND, either express or implied. See the License for the specific # language governing permissions and limitations under the License. from __future__ import absolute_import import datetime import io import logging import os import pytest import six from botocore.exceptions import ClientError from mock import ANY, MagicMock, Mock, patch, call, mock_open import sagemaker from sagemaker import TrainingInput, Session, get_execution_role from sagemaker.async_inference import AsyncInferenceConfig from sagemaker.session import ( _tuning_job_status, _transform_job_status, _train_done, NOTEBOOK_METADATA_FILE, ) from sagemaker.tuner import WarmStartConfig, WarmStartTypes from sagemaker.inputs import BatchDataCaptureConfig STATIC_HPs = {"feature_dim": "784"} SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}] ENV_INPUT = {"env_key1": "env_val1", "env_key2": "env_val2", "env_key3": "env_val3"} REGION = "us-west-2" STS_ENDPOINT = "sts.us-west-2.amazonaws.com" @pytest.fixture() def boto_session(): boto_mock = Mock(name="boto_session", region_name=REGION) client_mock = Mock() client_mock._client_config.user_agent = ( "Boto3/1.9.69 Python/3.6.5 Linux/4.14.77-70.82.amzn1.x86_64 Botocore/1.12.69 Resource" ) boto_mock.client.return_value = client_mock return boto_mock @patch("boto3.DEFAULT_SESSION") def test_default_session(boto3_default_session): sess = Session() assert sess.boto_session is boto3_default_session @patch("boto3.Session") def test_new_session_created(boto3_session): sess = Session() assert sess.boto_session is boto3_session.return_value def test_process(boto_session): session = Session(boto_session) process_request_args = { "inputs": [ { "InputName": "input-1", "S3Input": { "S3Uri": "mocked_s3_uri_from_upload_data", "LocalPath": "/container/path/", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, { "InputName": "my_dataset", "S3Input": { "S3Uri": "s3://path/to/my/dataset/census.csv", "LocalPath": "/container/path/", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, { "InputName": "source", "S3Input": { "S3Uri": "mocked_s3_uri_from_upload_data", "LocalPath": "/code/source", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, ], "output_config": { "Outputs": [ { "OutputName": "output-1", "S3Output": { "S3Uri": "s3://mybucket/current_job_name/output", "LocalPath": "/data/output", "S3UploadMode": "Continuous", }, }, { "OutputName": "my_output", "S3Output": { "S3Uri": "s3://uri/", "LocalPath": "/container/path/", "S3UploadMode": "Continuous", }, }, ], "KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key", }, "job_name": "current_job_name", "resources": { "ClusterConfig": { "InstanceType": "ml.m4.xlarge", "InstanceCount": 1, "VolumeSizeInGB": 100, } }, "stopping_condition": {"MaxRuntimeInSeconds": 3600}, "app_specification": { "ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3", "ContainerArguments": ["--drop-columns", "'SelfEmployed'"], "ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"], }, "environment": {"my_env_variable": 20}, "network_config": { "EnableInterContainerTrafficEncryption": True, "EnableNetworkIsolation": True, "VpcConfig": { "SecurityGroupIds": ["my_security_group_id"], "Subnets": ["my_subnet_id"], }, }, "role_arn": ROLE, "tags": [{"Name": "my-tag", "Value": "my-tag-value"}], "experiment_config": {"ExperimentName": "AnExperiment"}, } session.process(**process_request_args) expected_request = { "ProcessingJobName": "current_job_name", "ProcessingResources": { "ClusterConfig": { "InstanceType": "ml.m4.xlarge", "InstanceCount": 1, "VolumeSizeInGB": 100, } }, "AppSpecification": { "ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3", "ContainerArguments": ["--drop-columns", "'SelfEmployed'"], "ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"], }, "RoleArn": ROLE, "ProcessingInputs": [ { "InputName": "input-1", "S3Input": { "S3Uri": "mocked_s3_uri_from_upload_data", "LocalPath": "/container/path/", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, { "InputName": "my_dataset", "S3Input": { "S3Uri": "s3://path/to/my/dataset/census.csv", "LocalPath": "/container/path/", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, { "InputName": "source", "S3Input": { "S3Uri": "mocked_s3_uri_from_upload_data", "LocalPath": "/code/source", "S3DataType": "Archive", "S3InputMode": "File", "S3DataDistributionType": "FullyReplicated", "S3CompressionType": "None", }, }, ], "ProcessingOutputConfig": { "Outputs": [ { "OutputName": "output-1", "S3Output": { "S3Uri": "s3://mybucket/current_job_name/output", "LocalPath": "/data/output", "S3UploadMode": "Continuous", }, }, { "OutputName": "my_output", "S3Output": { "S3Uri": "s3://uri/", "LocalPath": "/container/path/", "S3UploadMode": "Continuous", }, }, ], "KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key", }, "Environment": {"my_env_variable": 20}, "NetworkConfig": { "EnableInterContainerTrafficEncryption": True, "EnableNetworkIsolation": True, "VpcConfig": { "SecurityGroupIds": ["my_security_group_id"], "Subnets": ["my_subnet_id"], }, }, "StoppingCondition": {"MaxRuntimeInSeconds": 3600}, "Tags": [{"Name": "my-tag", "Value": "my-tag-value"}], "ExperimentConfig": {"ExperimentName": "AnExperiment"}, } session.sagemaker_client.create_processing_job.assert_called_with(**expected_request) def mock_exists(filepath_to_mock, exists_result): unmocked_exists = os.path.exists def side_effect(filepath): if filepath == filepath_to_mock: return exists_result else: return unmocked_exists(filepath) return Mock(side_effect=side_effect) def test_get_execution_role(): session = Mock() session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:role/SageMakerRole" actual = get_execution_role(session) assert actual == "arn:aws:iam::369233609183:role/SageMakerRole" def test_get_execution_role_works_with_service_role(): session = Mock() session.get_caller_identity_arn.return_value = ( "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" ) actual = get_execution_role(session) assert ( actual == "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" ) def test_get_execution_role_throws_exception_if_arn_is_not_role(): session = Mock() session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos" with pytest.raises(ValueError) as error: get_execution_role(session) assert "The current AWS identity is not a role" in str(error.value) def test_get_execution_role_throws_exception_if_arn_is_not_role_with_role_in_name(): session = Mock() session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos-role" with pytest.raises(ValueError) as error: get_execution_role(session) assert "The current AWS identity is not a role" in str(error.value) @patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}')) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) def test_get_caller_identity_arn_from_describe_notebook_instance(boto_session): sess = Session(boto_session) expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" sess.sagemaker_client.describe_notebook_instance.return_value = {"RoleArn": expected_role} actual = sess.get_caller_identity_arn() assert actual == expected_role sess.sagemaker_client.describe_notebook_instance.assert_called_once_with( NotebookInstanceName="SageMakerInstance" ) @patch( "six.moves.builtins.open", mock_open( read_data='{"ResourceName": "SageMakerInstance", ' '"DomainId": "d-kbnw5yk6tg8j", ' '"UserProfileName": "default-1617915559064"}' ), ) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) def test_get_caller_identity_arn_from_describe_user_profile(boto_session): sess = Session(boto_session) expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" sess.sagemaker_client.describe_user_profile.return_value = { "UserSettings": {"ExecutionRole": expected_role} } actual = sess.get_caller_identity_arn() assert actual == expected_role sess.sagemaker_client.describe_user_profile.assert_called_once_with( DomainId="d-kbnw5yk6tg8j", UserProfileName="default-1617915559064", ) @patch( "six.moves.builtins.open", mock_open( read_data='{"ResourceName": "SageMakerInstance", ' '"DomainId": "d-kbnw5yk6tg8j", ' '"UserProfileName": "default-1617915559064"}' ), ) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) def test_get_caller_identity_arn_from_describe_domain_if_no_user_settings(boto_session): sess = Session(boto_session) expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388" sess.sagemaker_client.describe_user_profile.return_value = {} sess.sagemaker_client.describe_domain.return_value = { "DefaultUserSettings": {"ExecutionRole": expected_role} } actual = sess.get_caller_identity_arn() assert actual == expected_role sess.sagemaker_client.describe_user_profile.assert_called_once_with( DomainId="d-kbnw5yk6tg8j", UserProfileName="default-1617915559064", ) sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j") @patch( "six.moves.builtins.open", mock_open( read_data='{"ResourceName": "SageMakerInstance", ' '"DomainId": "d-kbnw5yk6tg8j", ' '"UserProfileName": "default-1617915559064"}' ), ) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) def test_fallback_to_domain_if_role_unavailable_in_user_settings(boto_session): sess = Session(boto_session) expected_role = "expected_role" sess.sagemaker_client.describe_user_profile.return_value = { "DomainId": "d-kbnw5yk6tg8j", "UserSettings": { "JupyterServerAppSettings": {}, "KernelGatewayAppSettings": {}, }, } sess.sagemaker_client.describe_domain.return_value = { "DefaultUserSettings": {"ExecutionRole": expected_role} } actual = sess.get_caller_identity_arn() assert actual == expected_role sess.sagemaker_client.describe_user_profile.assert_called_once_with( DomainId="d-kbnw5yk6tg8j", UserProfileName="default-1617915559064", ) sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j") @patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}')) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_a_role_after_describe_notebook_exception( sts_regional_endpoint, boto_session ): sess = Session(boto_session) exception = ClientError( {"Error": {"Code": "ValidationException", "Message": "RecordNotFound"}}, "Operation" ) sess.sagemaker_client.describe_notebook_instance.side_effect = exception arn = ( "arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122" ) sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": arn } expected_role = "arn:aws:iam::369233609183:role/SageMakerRole" sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}} with patch("logging.Logger.debug") as mock_logger: actual = sess.get_caller_identity_arn() mock_logger.assert_called_once() sess.sagemaker_client.describe_notebook_instance.assert_called_once_with( NotebookInstanceName="SageMakerInstance" ) assert actual == expected_role @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_a_user(sts_regional_endpoint, boto_session): sess = Session(boto_session) arn = "arn:aws:iam::369233609183:user/mia" sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": arn } sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": arn}} actual = sess.get_caller_identity_arn() assert actual == "arn:aws:iam::369233609183:user/mia" @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_an_user_without_permissions( sts_regional_endpoint, boto_session ): sess = Session(boto_session) arn = "arn:aws:iam::369233609183:user/mia" sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": arn } sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {}) with patch("logging.Logger.warning") as mock_logger: actual = sess.get_caller_identity_arn() assert actual == "arn:aws:iam::369233609183:user/mia" mock_logger.assert_called_once() @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_a_role(sts_regional_endpoint, boto_session): sess = Session(boto_session) arn = ( "arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122" ) sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": arn } expected_role = "arn:aws:iam::369233609183:role/SageMakerRole" sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}} actual = sess.get_caller_identity_arn() assert actual == expected_role @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_an_execution_role(sts_regional_endpoint, boto_session): sess = Session(boto_session) sts_arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker" sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": sts_arn } iam_arn = "arn:aws:iam::369233609183:role/AmazonSageMaker-ExecutionRole-20171129T072388" sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": iam_arn}} actual = sess.get_caller_identity_arn() assert actual == iam_arn @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_a_sagemaker_execution_role_with_iam_client_error( sts_regional_endpoint, boto_session ): sess = Session(boto_session) arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker" sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": arn } sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {}) actual = sess.get_caller_identity_arn() assert ( actual == "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388" ) @patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False)) @patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT) def test_get_caller_identity_arn_from_role_with_path(sts_regional_endpoint, boto_session): sess = Session(boto_session) arn_prefix = "arn:aws:iam::369233609183:role" role_name = "name" sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Arn": "/".join([arn_prefix, role_name]) } role_path = "path" role_with_path = "/".join([arn_prefix, role_path, role_name]) sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": role_with_path}} actual = sess.get_caller_identity_arn() assert actual == role_with_path def test_delete_endpoint(boto_session): sess = Session(boto_session) sess.delete_endpoint("my_endpoint") boto_session.client().delete_endpoint.assert_called_with(EndpointName="my_endpoint") def test_delete_endpoint_config(boto_session): sess = Session(boto_session) sess.delete_endpoint_config("my_endpoint_config") boto_session.client().delete_endpoint_config.assert_called_with( EndpointConfigName="my_endpoint_config" ) def test_delete_model(boto_session): sess = Session(boto_session) model_name = "my_model" sess.delete_model(model_name) boto_session.client().delete_model.assert_called_with(ModelName=model_name) def test_user_agent_injected(boto_session): assert ( "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent ) sess = Session(boto_session) assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent assert ( "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_runtime_client._client_config.user_agent ) def test_user_agent_injected_with_nbi(boto_session): assert ( "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent ) with patch("six.moves.builtins.open", mock_open(read_data="120.0-0")) as mo: sess = Session(boto_session) mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt") assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent assert "AWS-SageMaker-Notebook-Instance" in sess.sagemaker_client._client_config.user_agent assert ( "AWS-SageMaker-Notebook-Instance" in sess.sagemaker_runtime_client._client_config.user_agent ) def test_user_agent_injected_with_nbi_ioerror(boto_session): assert ( "AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent ) with patch("six.moves.builtins.open", MagicMock(side_effect=IOError("File not found"))) as mo: sess = Session(boto_session) mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt") assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent assert ( "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_runtime_client._client_config.user_agent ) def test_training_input_all_defaults(): prefix = "pre" actual = TrainingInput(s3_data=prefix) expected = { "DataSource": { "S3DataSource": { "S3DataDistributionType": "FullyReplicated", "S3DataType": "S3Prefix", "S3Uri": prefix, } } } assert actual.config == expected def test_training_input_all_arguments(): prefix = "pre" distribution = "FullyReplicated" compression = "Gzip" content_type = "text/csv" record_wrapping = "RecordIO" s3_data_type = "Manifestfile" input_mode = "Pipe" result = TrainingInput( s3_data=prefix, distribution=distribution, compression=compression, input_mode=input_mode, content_type=content_type, record_wrapping=record_wrapping, s3_data_type=s3_data_type, ) expected = { "DataSource": { "S3DataSource": { "S3DataDistributionType": distribution, "S3DataType": s3_data_type, "S3Uri": prefix, } }, "CompressionType": compression, "ContentType": content_type, "RecordWrapperType": record_wrapping, "InputMode": input_mode, } assert result.config == expected IMAGE = "myimage" S3_INPUT_URI = "s3://mybucket/data" S3_OUTPUT = "s3://sagemaker-123/output/jobname" ROLE = "SageMakerRole" EXPANDED_ROLE = "arn:aws:iam::111111111111:role/ExpandedRole" INSTANCE_COUNT = 1 INSTANCE_TYPE = "ml.c4.xlarge" ACCELERATOR_TYPE = "ml.eia.medium" MAX_SIZE = 30 MAX_TIME = 3 * 60 * 60 JOB_NAME = "jobname" TAGS = [{"Name": "some-tag", "Value": "value-for-tag"}] VPC_CONFIG = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]} METRIC_DEFINITONS = [{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}] EXPERIMENT_CONFIG = { "ExperimentName": "dummyExp", "TrialName": "dummyT", "TrialComponentDisplayName": "dummyTC", } MODEL_CLIENT_CONFIG = {"InvocationsMaxRetries": 2, "InvocationsTimeoutInSeconds": 60} DEFAULT_EXPECTED_TRAIN_JOB_ARGS = { "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, "RoleArn": EXPANDED_ROLE, "ResourceConfig": { "InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE, "VolumeSizeInGB": MAX_SIZE, }, "InputDataConfig": [ { "DataSource": { "S3DataSource": { "S3DataDistributionType": "FullyReplicated", "S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI, } }, "ChannelName": "training", } ], "AlgorithmSpecification": {"TrainingInputMode": "File", "TrainingImage": IMAGE}, "TrainingJobName": JOB_NAME, "StoppingCondition": {"MaxRuntimeInSeconds": MAX_TIME}, "VpcConfig": VPC_CONFIG, "ExperimentConfig": EXPERIMENT_CONFIG, } COMPLETED_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS) COMPLETED_DESCRIBE_JOB_RESULT.update( {"TrainingJobArn": "arn:aws:sagemaker:us-west-2:336:training-job/" + JOB_NAME} ) COMPLETED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Completed"}) COMPLETED_DESCRIBE_JOB_RESULT.update( {"ModelArtifacts": {"S3ModelArtifacts": S3_OUTPUT + "/model/model.tar.gz"}} ) # TrainingStartTime and TrainingEndTime are for billable seconds calculation COMPLETED_DESCRIBE_JOB_RESULT.update( {"TrainingStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000)} ) COMPLETED_DESCRIBE_JOB_RESULT.update( {"TrainingEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000)} ) STOPPED_DESCRIBE_JOB_RESULT = dict(COMPLETED_DESCRIBE_JOB_RESULT) STOPPED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Stopped"}) IN_PROGRESS_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS) IN_PROGRESS_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "InProgress"}) COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT = { "TransformJobStatus": "Completed", "ModelName": "some-model", "TransformJobName": JOB_NAME, "TransformResources": {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE}, "TransformEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000), "TransformStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000), "TransformOutput": {"AssembleWith": "None", "KmsKeyId": "", "S3OutputPath": S3_OUTPUT}, "TransformInput": { "CompressionType": "None", "ContentType": "text/csv", "DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}, "SplitType": "Line", }, } STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT) STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "Stopped"}) IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT) IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "InProgress"}) SERVERLESS_INFERENCE_CONFIG = { "MemorySizeInMB": 2048, "MaxConcurrency": 2, } @pytest.fixture() def sagemaker_session(): boto_mock = MagicMock(name="boto_session") boto_mock.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = { "Account": "123" } ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) ims.expand_role = Mock(return_value=EXPANDED_ROLE) return ims def test_train_pack_to_request(sagemaker_session): in_config = [ { "ChannelName": "training", "DataSource": { "S3DataSource": { "S3DataDistributionType": "FullyReplicated", "S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI, } }, } ] out_config = {"S3OutputPath": S3_OUTPUT} resource_config = { "InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE, "VolumeSizeInGB": MAX_SIZE, } stop_cond = {"MaxRuntimeInSeconds": MAX_TIME} sagemaker_session.train( image_uri=IMAGE, input_mode="File", input_config=in_config, role=EXPANDED_ROLE, job_name=JOB_NAME, output_config=out_config, resource_config=resource_config, hyperparameters=None, stop_condition=stop_cond, tags=None, vpc_config=VPC_CONFIG, metric_definitions=None, experiment_config=EXPERIMENT_CONFIG, enable_sagemaker_metrics=None, ) assert sagemaker_session.sagemaker_client.method_calls[0] == ( "create_training_job", (), DEFAULT_EXPECTED_TRAIN_JOB_ARGS, ) SAMPLE_STOPPING_CONDITION = {"MaxRuntimeInSeconds": MAX_TIME} RESOURCE_CONFIG = { "InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE, "VolumeSizeInGB": MAX_SIZE, } SAMPLE_INPUT = [ { "DataSource": { "S3DataSource": { "S3DataDistributionType": "FullyReplicated", "S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI, } }, "ChannelName": "training", } ] SAMPLE_OUTPUT = {"S3OutputPath": S3_OUTPUT} SAMPLE_OBJECTIVE = {"Type": "Maximize", "MetricName": "val-score"} SAMPLE_OBJECTIVE_2 = {"Type": "Maximize", "MetricName": "value-score"} SAMPLE_METRIC_DEF = [{"Name": "train:progress", "Regex": "regex-1"}] SAMPLE_METRIC_DEF_2 = [{"Name": "value-score", "Regex": "regex-2"}] STATIC_HPs = {"feature_dim": "784"} STATIC_HPs_2 = {"gamma": "0.1"} SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}] SAMPLE_PARAM_RANGES_2 = [{"Name": "kernel", "Values": ["rbf", "sigmoid"]}] SAMPLE_TUNING_JOB_REQUEST = { "HyperParameterTuningJobName": "dummy-tuning-1", "HyperParameterTuningJobConfig": { "Strategy": "Bayesian", "HyperParameterTuningJobObjective": SAMPLE_OBJECTIVE, "ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5}, "ParameterRanges": SAMPLE_PARAM_RANGES, "TrainingJobEarlyStoppingType": "Off", }, "TrainingJobDefinition": { "StaticHyperParameters": STATIC_HPs, "AlgorithmSpecification": { "TrainingImage": "dummy-image-1", "TrainingInputMode": "File", "MetricDefinitions": SAMPLE_METRIC_DEF, }, "RoleArn": EXPANDED_ROLE, "InputDataConfig": SAMPLE_INPUT, "OutputDataConfig": SAMPLE_OUTPUT, "ResourceConfig": RESOURCE_CONFIG, "StoppingCondition": SAMPLE_STOPPING_CONDITION, }, } SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST = { "HyperParameterTuningJobName": "dummy-tuning-1", "HyperParameterTuningJobConfig": { "Strategy": "Bayesian", "ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5}, "TrainingJobEarlyStoppingType": "Off", }, "TrainingJobDefinitions": [ { "DefinitionName": "estimator_1", "TuningObjective": SAMPLE_OBJECTIVE, "HyperParameterRanges": SAMPLE_PARAM_RANGES, "StaticHyperParameters": STATIC_HPs, "AlgorithmSpecification": { "TrainingImage": "dummy-image-1", "TrainingInputMode": "File", "MetricDefinitions": SAMPLE_METRIC_DEF, }, "RoleArn": EXPANDED_ROLE, "InputDataConfig": SAMPLE_INPUT, "OutputDataConfig": SAMPLE_OUTPUT, "ResourceConfig": RESOURCE_CONFIG, "StoppingCondition": SAMPLE_STOPPING_CONDITION, }, { "DefinitionName": "estimator_2", "TuningObjective": SAMPLE_OBJECTIVE_2, "HyperParameterRanges": SAMPLE_PARAM_RANGES_2, "StaticHyperParameters": STATIC_HPs_2, "AlgorithmSpecification": { "TrainingImage": "dummy-image-2", "TrainingInputMode": "File", "MetricDefinitions": SAMPLE_METRIC_DEF_2, }, "RoleArn": EXPANDED_ROLE, "InputDataConfig": SAMPLE_INPUT, "OutputDataConfig": SAMPLE_OUTPUT, "ResourceConfig": RESOURCE_CONFIG, "StoppingCondition": SAMPLE_STOPPING_CONDITION, }, ], } @pytest.mark.parametrize( "warm_start_type, parents", [("IdenticalDataAndAlgorithm", {"p1", "p2", "p3"}), ("TransferLearning", {"p1", "p2", "p3"})], ) def test_tune_warm_start(sagemaker_session, warm_start_type, parents): def assert_create_tuning_job_request(**kwrags): assert ( kwrags["HyperParameterTuningJobConfig"] == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] ) assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] assert kwrags["WarmStartConfig"] == { "WarmStartType": warm_start_type, "ParentHyperParameterTuningJobs": [ {"HyperParameterTuningJobName": parent} for parent in parents ], } sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.tune( job_name="dummy-tuning-1", strategy="Bayesian", objective_type="Maximize", objective_metric_name="val-score", max_jobs=100, max_parallel_jobs=5, parameter_ranges=SAMPLE_PARAM_RANGES, static_hyperparameters=STATIC_HPs, image_uri="dummy-image-1", input_mode="File", metric_definitions=SAMPLE_METRIC_DEF, role=EXPANDED_ROLE, input_config=SAMPLE_INPUT, output_config=SAMPLE_OUTPUT, resource_config=RESOURCE_CONFIG, stop_condition=SAMPLE_STOPPING_CONDITION, tags=None, warm_start_config=WarmStartConfig( warm_start_type=WarmStartTypes(warm_start_type), parents=parents ).to_input_req(), ) def test_create_tuning_job_without_training_config_or_list(sagemaker_session): with pytest.raises( ValueError, match="Either training_config or training_config_list should be provided." ): sagemaker_session.create_tuning_job( job_name="dummy-tuning-1", tuning_config={ "strategy": "Bayesian", "objective_type": "Maximize", "objective_metric_name": "val-score", "max_jobs": 100, "max_parallel_jobs": 5, "parameter_ranges": SAMPLE_PARAM_RANGES, }, ) def test_create_tuning_job_with_both_training_config_and_list(sagemaker_session): with pytest.raises( ValueError, match="Only one of training_config and training_config_list should be provided." ): sagemaker_session.create_tuning_job( job_name="dummy-tuning-1", tuning_config={ "strategy": "Bayesian", "objective_type": "Maximize", "objective_metric_name": "val-score", "max_jobs": 100, "max_parallel_jobs": 5, "parameter_ranges": SAMPLE_PARAM_RANGES, }, training_config={"static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1"}, training_config_list=[ { "static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1", "estimator_name": "estimator_1", }, { "static_hyperparameters": STATIC_HPs_2, "image_uri": "dummy-image-2", "estimator_name": "estimator_2", }, ], ) def test_create_tuning_job(sagemaker_session): def assert_create_tuning_job_request(**kwrags): assert ( kwrags["HyperParameterTuningJobConfig"] == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] ) assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] assert "TrainingJobDefinitions" not in kwrags assert kwrags.get("WarmStartConfig", None) is None sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.create_tuning_job( job_name="dummy-tuning-1", tuning_config={ "strategy": "Bayesian", "objective_type": "Maximize", "objective_metric_name": "val-score", "max_jobs": 100, "max_parallel_jobs": 5, "parameter_ranges": SAMPLE_PARAM_RANGES, }, training_config={ "static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1", "input_mode": "File", "metric_definitions": SAMPLE_METRIC_DEF, "role": EXPANDED_ROLE, "input_config": SAMPLE_INPUT, "output_config": SAMPLE_OUTPUT, "resource_config": RESOURCE_CONFIG, "stop_condition": SAMPLE_STOPPING_CONDITION, }, tags=None, warm_start_config=None, ) def test_create_tuning_job_multi_algo(sagemaker_session): def assert_create_tuning_job_request(**kwrags): expected_tuning_config = SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST[ "HyperParameterTuningJobConfig" ] assert kwrags["HyperParameterTuningJobConfig"] == expected_tuning_config assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" assert "TrainingJobDefinition" not in kwrags assert ( kwrags["TrainingJobDefinitions"] == SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST["TrainingJobDefinitions"] ) assert kwrags.get("WarmStartConfig", None) is None sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.create_tuning_job( job_name="dummy-tuning-1", tuning_config={"strategy": "Bayesian", "max_jobs": 100, "max_parallel_jobs": 5}, training_config_list=[ { "static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1", "input_mode": "File", "metric_definitions": SAMPLE_METRIC_DEF, "role": EXPANDED_ROLE, "input_config": SAMPLE_INPUT, "output_config": SAMPLE_OUTPUT, "resource_config": RESOURCE_CONFIG, "stop_condition": SAMPLE_STOPPING_CONDITION, "estimator_name": "estimator_1", "objective_type": "Maximize", "objective_metric_name": "val-score", "parameter_ranges": SAMPLE_PARAM_RANGES, }, { "static_hyperparameters": STATIC_HPs_2, "image_uri": "dummy-image-2", "input_mode": "File", "metric_definitions": SAMPLE_METRIC_DEF_2, "role": EXPANDED_ROLE, "input_config": SAMPLE_INPUT, "output_config": SAMPLE_OUTPUT, "resource_config": RESOURCE_CONFIG, "stop_condition": SAMPLE_STOPPING_CONDITION, "estimator_name": "estimator_2", "objective_type": "Maximize", "objective_metric_name": "value-score", "parameter_ranges": SAMPLE_PARAM_RANGES_2, }, ], tags=None, warm_start_config=None, ) def test_tune(sagemaker_session): def assert_create_tuning_job_request(**kwrags): assert ( kwrags["HyperParameterTuningJobConfig"] == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] ) assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"] assert kwrags.get("WarmStartConfig", None) is None sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.tune( job_name="dummy-tuning-1", strategy="Bayesian", objective_type="Maximize", objective_metric_name="val-score", max_jobs=100, max_parallel_jobs=5, parameter_ranges=SAMPLE_PARAM_RANGES, static_hyperparameters=STATIC_HPs, image_uri="dummy-image-1", input_mode="File", metric_definitions=SAMPLE_METRIC_DEF, role=EXPANDED_ROLE, input_config=SAMPLE_INPUT, output_config=SAMPLE_OUTPUT, resource_config=RESOURCE_CONFIG, stop_condition=SAMPLE_STOPPING_CONDITION, tags=None, warm_start_config=None, ) def test_tune_with_encryption_flag(sagemaker_session): def assert_create_tuning_job_request(**kwrags): assert ( kwrags["HyperParameterTuningJobConfig"] == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] ) assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1" assert kwrags["TrainingJobDefinition"]["EnableInterContainerTrafficEncryption"] is True assert kwrags.get("WarmStartConfig", None) is None sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.tune( job_name="dummy-tuning-1", strategy="Bayesian", objective_type="Maximize", objective_metric_name="val-score", max_jobs=100, max_parallel_jobs=5, parameter_ranges=SAMPLE_PARAM_RANGES, static_hyperparameters=STATIC_HPs, image_uri="dummy-image-1", input_mode="File", metric_definitions=SAMPLE_METRIC_DEF, role=EXPANDED_ROLE, input_config=SAMPLE_INPUT, output_config=SAMPLE_OUTPUT, resource_config=RESOURCE_CONFIG, stop_condition=SAMPLE_STOPPING_CONDITION, tags=None, warm_start_config=None, encrypt_inter_container_traffic=True, ) def test_tune_with_spot_and_checkpoints(sagemaker_session): def assert_create_tuning_job_request(**kwargs): assert ( kwargs["HyperParameterTuningJobConfig"] == SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"] ) assert kwargs["HyperParameterTuningJobName"] == "dummy-tuning-1" assert kwargs["TrainingJobDefinition"]["EnableManagedSpotTraining"] is True assert ( kwargs["TrainingJobDefinition"]["CheckpointConfig"]["S3Uri"] == "s3://mybucket/checkpoints/" ) assert ( kwargs["TrainingJobDefinition"]["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints" ) assert kwargs.get("WarmStartConfig", None) is None sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = ( assert_create_tuning_job_request ) sagemaker_session.tune( job_name="dummy-tuning-1", strategy="Bayesian", objective_type="Maximize", objective_metric_name="val-score", max_jobs=100, max_parallel_jobs=5, parameter_ranges=SAMPLE_PARAM_RANGES, static_hyperparameters=STATIC_HPs, image_uri="dummy-image-1", input_mode="File", metric_definitions=SAMPLE_METRIC_DEF, role=EXPANDED_ROLE, input_config=SAMPLE_INPUT, output_config=SAMPLE_OUTPUT, resource_config=RESOURCE_CONFIG, stop_condition=SAMPLE_STOPPING_CONDITION, tags=None, warm_start_config=None, use_spot_instances=True, checkpoint_s3_uri="s3://mybucket/checkpoints/", checkpoint_local_path="/tmp/checkpoints", ) def test_stop_tuning_job(sagemaker_session): sms = sagemaker_session sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( name="stop_hyper_parameter_tuning_job" ) sagemaker_session.stop_tuning_job(JOB_NAME) sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( HyperParameterTuningJobName=JOB_NAME ) def test_stop_tuning_job_client_error_already_stopped(sagemaker_session): sms = sagemaker_session exception = ClientError({"Error": {"Code": "ValidationException"}}, "Operation") sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( name="stop_hyper_parameter_tuning_job", side_effect=exception ) sagemaker_session.stop_tuning_job(JOB_NAME) sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( HyperParameterTuningJobName=JOB_NAME ) def test_stop_tuning_job_client_error(sagemaker_session): error_response = {"Error": {"Code": "MockException", "Message": "MockMessage"}} operation = "Operation" exception = ClientError(error_response, operation) sms = sagemaker_session sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock( name="stop_hyper_parameter_tuning_job", side_effect=exception ) with pytest.raises(ClientError) as e: sagemaker_session.stop_tuning_job(JOB_NAME) sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with( HyperParameterTuningJobName=JOB_NAME ) assert ( "An error occurred (MockException) when calling the Operation operation: MockMessage" in str(e) ) def test_train_pack_to_request_with_optional_params(sagemaker_session): in_config = [ { "ChannelName": "training", "DataSource": { "S3DataSource": { "S3DataDistributionType": "FullyReplicated", "S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI, } }, } ] out_config = {"S3OutputPath": S3_OUTPUT} resource_config = { "InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE, "VolumeSizeInGB": MAX_SIZE, } stop_cond = {"MaxRuntimeInSeconds": MAX_TIME} RETRY_STRATEGY = {"MaximumRetryAttempts": 2} hyperparameters = {"foo": "bar"} sagemaker_session.train( image_uri=IMAGE, input_mode="File", input_config=in_config, role=EXPANDED_ROLE, job_name=JOB_NAME, output_config=out_config, resource_config=resource_config, vpc_config=VPC_CONFIG, hyperparameters=hyperparameters, stop_condition=stop_cond, tags=TAGS, metric_definitions=METRIC_DEFINITONS, encrypt_inter_container_traffic=True, use_spot_instances=True, checkpoint_s3_uri="s3://mybucket/checkpoints/", checkpoint_local_path="/tmp/checkpoints", enable_sagemaker_metrics=True, environment=ENV_INPUT, retry_strategy=RETRY_STRATEGY, ) _, _, actual_train_args = sagemaker_session.sagemaker_client.method_calls[0] assert actual_train_args["VpcConfig"] == VPC_CONFIG assert actual_train_args["HyperParameters"] == hyperparameters assert actual_train_args["Tags"] == TAGS assert actual_train_args["AlgorithmSpecification"]["MetricDefinitions"] == METRIC_DEFINITONS assert actual_train_args["AlgorithmSpecification"]["EnableSageMakerMetricsTimeSeries"] is True assert actual_train_args["EnableInterContainerTrafficEncryption"] is True assert actual_train_args["EnableManagedSpotTraining"] is True assert actual_train_args["CheckpointConfig"]["S3Uri"] == "s3://mybucket/checkpoints/" assert actual_train_args["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints" assert actual_train_args["Environment"] == ENV_INPUT assert actual_train_args["RetryStrategy"] == RETRY_STRATEGY def test_transform_pack_to_request(sagemaker_session): model_name = "my-model" in_config = { "CompressionType": "None", "ContentType": "text/csv", "SplitType": "None", "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, } out_config = {"S3OutputPath": S3_OUTPUT} resource_config = {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE} data_processing = {"OutputFilter": "$", "InputFilter": "$", "JoinSource": "Input"} expected_args = { "TransformJobName": JOB_NAME, "ModelName": model_name, "TransformInput": in_config, "TransformOutput": out_config, "TransformResources": resource_config, "DataProcessing": data_processing, } sagemaker_session.transform( job_name=JOB_NAME, model_name=model_name, strategy=None, max_concurrent_transforms=None, max_payload=None, env=None, input_config=in_config, output_config=out_config, resource_config=resource_config, experiment_config=None, model_client_config=None, tags=None, data_processing=data_processing, batch_data_capture_config=None, ) _, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0] assert actual_args == expected_args def test_transform_pack_to_request_with_optional_params(sagemaker_session): strategy = "strategy" max_concurrent_transforms = 1 max_payload = 0 env = {"FOO": "BAR"} batch_data_capture_config = BatchDataCaptureConfig( destination_s3_uri="test_uri", kms_key_id="", generate_inference_id=False, ) sagemaker_session.transform( job_name=JOB_NAME, model_name="my-model", strategy=strategy, max_concurrent_transforms=max_concurrent_transforms, env=env, max_payload=max_payload, input_config={}, output_config={}, resource_config={}, experiment_config=EXPERIMENT_CONFIG, model_client_config=MODEL_CLIENT_CONFIG, tags=TAGS, data_processing=None, batch_data_capture_config=batch_data_capture_config, ) _, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0] assert actual_args["BatchStrategy"] == strategy assert actual_args["MaxConcurrentTransforms"] == max_concurrent_transforms assert actual_args["MaxPayloadInMB"] == max_payload assert actual_args["Environment"] == env assert actual_args["Tags"] == TAGS assert actual_args["ExperimentConfig"] == EXPERIMENT_CONFIG assert actual_args["ModelClientConfig"] == MODEL_CLIENT_CONFIG assert actual_args["DataCaptureConfig"] == batch_data_capture_config._to_request_dict() @patch("sys.stdout", new_callable=io.BytesIO if six.PY2 else io.StringIO) def test_color_wrap(bio): color_wrap = sagemaker.logs.ColorWrap() color_wrap(0, "hi there") assert bio.getvalue() == "hi there\n" class MockBotoException(ClientError): def __init__(self, code): self.response = {"Error": {"Code": code}} DEFAULT_LOG_STREAMS = {"logStreams": [{"logStreamName": JOB_NAME + "/xxxxxxxxx"}]} LIFECYCLE_LOG_STREAMS = [ MockBotoException("ResourceNotFoundException"), DEFAULT_LOG_STREAMS, DEFAULT_LOG_STREAMS, DEFAULT_LOG_STREAMS, DEFAULT_LOG_STREAMS, DEFAULT_LOG_STREAMS, DEFAULT_LOG_STREAMS, ] DEFAULT_LOG_EVENTS = [ {"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]}, {"nextForwardToken": None, "events": []}, ] STREAM_LOG_EVENTS = [ {"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]}, {"nextForwardToken": None, "events": []}, { "nextForwardToken": None, "events": [ {"timestamp": 1, "message": "hi there #1"}, {"timestamp": 2, "message": "hi there #2"}, ], }, {"nextForwardToken": None, "events": []}, { "nextForwardToken": None, "events": [ {"timestamp": 2, "message": "hi there #2"}, {"timestamp": 2, "message": "hi there #2a"}, {"timestamp": 3, "message": "hi there #3"}, ], }, {"nextForwardToken": None, "events": []}, ] @pytest.fixture() def sagemaker_session_complete(): boto_mock = MagicMock(name="boto_session") boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.sagemaker_client.describe_transform_job.return_value = ( COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT ) return ims @pytest.fixture() def sagemaker_session_stopped(): boto_mock = MagicMock(name="boto_session") boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) ims.sagemaker_client.describe_training_job.return_value = STOPPED_DESCRIBE_JOB_RESULT ims.sagemaker_client.describe_transform_job.return_value = STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT return ims @pytest.fixture() def sagemaker_session_ready_lifecycle(): boto_mock = MagicMock(name="boto_session") boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) ims.sagemaker_client.describe_training_job.side_effect = [ IN_PROGRESS_DESCRIBE_JOB_RESULT, IN_PROGRESS_DESCRIBE_JOB_RESULT, COMPLETED_DESCRIBE_JOB_RESULT, ] ims.sagemaker_client.describe_transform_job.side_effect = [ IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT, ] return ims @pytest.fixture() def sagemaker_session_full_lifecycle(): boto_mock = MagicMock(name="boto_session") boto_mock.client("logs").describe_log_streams.side_effect = LIFECYCLE_LOG_STREAMS boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock()) ims.sagemaker_client.describe_training_job.side_effect = [ IN_PROGRESS_DESCRIBE_JOB_RESULT, IN_PROGRESS_DESCRIBE_JOB_RESULT, COMPLETED_DESCRIBE_JOB_RESULT, ] ims.sagemaker_client.describe_transform_job.side_effect = [ IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT, COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT, ] return ims @patch("sagemaker.logs.ColorWrap") def test_logs_for_job_no_wait(cw, sagemaker_session_complete): ims = sagemaker_session_complete ims.logs_for_job(JOB_NAME) ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME) cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_job_no_wait_stopped_job(cw, sagemaker_session_stopped): ims = sagemaker_session_stopped ims.logs_for_job(JOB_NAME) ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME) cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_job_wait_on_completed(cw, sagemaker_session_complete): ims = sagemaker_session_complete ims.logs_for_job(JOB_NAME, wait=True, poll=0) assert ims.sagemaker_client.describe_training_job.call_args_list == [ call(TrainingJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_job_wait_on_stopped(cw, sagemaker_session_stopped): ims = sagemaker_session_stopped ims.logs_for_job(JOB_NAME, wait=True, poll=0) assert ims.sagemaker_client.describe_training_job.call_args_list == [ call(TrainingJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle): ims = sagemaker_session_ready_lifecycle ims.logs_for_job(JOB_NAME) assert ims.sagemaker_client.describe_training_job.call_args_list == [ call(TrainingJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") @patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180]) def test_logs_for_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle): ims = sagemaker_session_full_lifecycle ims.logs_for_job(JOB_NAME, wait=True, poll=0) assert ( ims.sagemaker_client.describe_training_job.call_args_list == [call(TrainingJobName=JOB_NAME)] * 3 ) assert cw().call_args_list == [ call(0, "hi there #1"), call(0, "hi there #2"), call(0, "hi there #2a"), call(0, "hi there #3"), ] @patch("sagemaker.logs.ColorWrap") def test_logs_for_transform_job_no_wait(cw, sagemaker_session_complete): ims = sagemaker_session_complete ims.logs_for_transform_job(JOB_NAME) ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME) cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_transform_job_no_wait_stopped_job(cw, sagemaker_session_stopped): ims = sagemaker_session_stopped ims.logs_for_transform_job(JOB_NAME) ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME) cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_transform_job_wait_on_completed(cw, sagemaker_session_complete): ims = sagemaker_session_complete ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) assert ims.sagemaker_client.describe_transform_job.call_args_list == [ call(TransformJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_transform_job_wait_on_stopped(cw, sagemaker_session_stopped): ims = sagemaker_session_stopped ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) assert ims.sagemaker_client.describe_transform_job.call_args_list == [ call(TransformJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") def test_logs_for_transform_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle): ims = sagemaker_session_ready_lifecycle ims.logs_for_transform_job(JOB_NAME) assert ims.sagemaker_client.describe_transform_job.call_args_list == [ call(TransformJobName=JOB_NAME) ] cw().assert_called_with(0, "hi there #1") @patch("sagemaker.logs.ColorWrap") @patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180]) def test_logs_for_transform_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle): ims = sagemaker_session_full_lifecycle ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0) assert ( ims.sagemaker_client.describe_transform_job.call_args_list == [call(TransformJobName=JOB_NAME)] * 3 ) assert cw().call_args_list == [ call(0, "hi there #1"), call(0, "hi there #2"), call(0, "hi there #2a"), call(0, "hi there #3"), ] MODEL_NAME = "some-model" PRIMARY_CONTAINER = { "Environment": {}, "Image": IMAGE, "ModelDataUrl": "s3://sagemaker-123/output/jobname/model/model.tar.gz", } @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model(expand_container_def, sagemaker_session): model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) assert model == MODEL_NAME sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_with_tags(expand_container_def, sagemaker_session): tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}] model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, tags=tags) assert model == MODEL_NAME tags = [{"Value": "TagtestValue", "Key": "TagtestKey"}] sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER, Tags=tags, ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_with_primary_container(expand_container_def, sagemaker_session): model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER) assert model == MODEL_NAME sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_with_both(expand_container_def, sagemaker_session): with pytest.raises(ValueError): sagemaker_session.create_model( MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER, primary_container=PRIMARY_CONTAINER ) CONTAINERS = [ { "Environment": {"SAGEMAKER_DEFAULT_INVOCATIONS_ACCEPT": "application/json"}, "Image": "mi-1", "ModelDataUrl": "s3://bucket/model_1.tar.gz", "Framework": "TENSORFLOW", "FrameworkVersion": "2.9", "NearestModelName": "resnet50", "ModelInput": { "DataInputConfig": '{"input_1":[1,224,224,3]}', }, }, {"Environment": {}, "Image": "mi-2", "ModelDataUrl": "s3://bucket/model_2.tar.gz"}, ] @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_pipeline_model(expand_container_def, sagemaker_session): model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=CONTAINERS) assert model == MODEL_NAME sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, Containers=CONTAINERS ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_vpc_config(expand_container_def, sagemaker_session): model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, VPC_CONFIG) assert model == MODEL_NAME sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER, VpcConfig=VPC_CONFIG, ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_pipeline_model_vpc_config(expand_container_def, sagemaker_session): model = sagemaker_session.create_model(MODEL_NAME, ROLE, CONTAINERS, VPC_CONFIG) assert model == MODEL_NAME sagemaker_session.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, Containers=CONTAINERS, VpcConfig=VPC_CONFIG, ) @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_already_exists(expand_container_def, sagemaker_session, caplog): error_response = { "Error": {"Code": "ValidationException", "Message": "Cannot create already existing model"} } exception = ClientError(error_response, "Operation") sagemaker_session.sagemaker_client.create_model.side_effect = exception model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) assert model == MODEL_NAME expected_warning = ( "sagemaker", logging.WARNING, "Using already existing model: {}".format(MODEL_NAME), ) assert expected_warning in caplog.record_tuples @patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER) def test_create_model_failure(expand_container_def, sagemaker_session): error_message = "this is expected" sagemaker_session.sagemaker_client.create_model.side_effect = RuntimeError(error_message) with pytest.raises(RuntimeError) as e: sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER) assert error_message in str(e) def test_create_model_from_job(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.create_model_from_job(JOB_NAME) assert ( call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list ) ims.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=JOB_NAME, PrimaryContainer=PRIMARY_CONTAINER, VpcConfig=VPC_CONFIG, ) def test_create_model_from_job_with_tags(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.create_model_from_job(JOB_NAME, tags=TAGS) assert ( call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list ) ims.sagemaker_client.create_model.assert_called_with( ExecutionRoleArn=EXPANDED_ROLE, ModelName=JOB_NAME, PrimaryContainer=PRIMARY_CONTAINER, VpcConfig=VPC_CONFIG, Tags=TAGS, ) def test_create_model_from_job_with_image(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.create_model_from_job(JOB_NAME, image_uri="some-image") [create_model_call] = ims.sagemaker_client.create_model.call_args_list assert dict(create_model_call[1]["PrimaryContainer"])["Image"] == "some-image" def test_create_model_from_job_with_container_def(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.create_model_from_job( JOB_NAME, image_uri="some-image", model_data_url="some-data", env={"a": "b"}, ) [create_model_call] = ims.sagemaker_client.create_model.call_args_list c_def = create_model_call[1]["PrimaryContainer"] assert c_def["Image"] == "some-image" assert c_def["ModelDataUrl"] == "some-data" assert c_def["Environment"] == {"a": "b"} def test_create_model_from_job_with_vpc_config_override(sagemaker_session): vpc_config_override = {"Subnets": ["foo", "bar"], "SecurityGroupIds": ["baz"]} ims = sagemaker_session ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT ims.create_model_from_job(JOB_NAME, vpc_config_override=vpc_config_override) assert ims.sagemaker_client.create_model.call_args[1]["VpcConfig"] == vpc_config_override ims.create_model_from_job(JOB_NAME, vpc_config_override=None) assert "VpcConfig" not in ims.sagemaker_client.create_model.call_args[1] def test_endpoint_from_production_variants(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) pvs = [ sagemaker.production_variant("A", "ml.p2.xlarge"), sagemaker.production_variant("B", "p299.4096xlarge"), ] ex = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" ) ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs) sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[] ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="some-endpoint", ProductionVariants=pvs ) def test_create_endpoint_config_with_tags(sagemaker_session): tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}] sagemaker_session.create_endpoint_config("endpoint-test", "simple-model", 1, "local", tags=tags) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="endpoint-test", ProductionVariants=ANY, Tags=tags ) def test_endpoint_from_production_variants_with_tags(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) pvs = [ sagemaker.production_variant("A", "ml.p2.xlarge"), sagemaker.production_variant("B", "p299.4096xlarge"), ] ex = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" ) ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) tags = [{"ModelName": "TestModel"}] sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags ) def test_endpoint_from_production_variants_with_accelerator_type(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) pvs = [ sagemaker.production_variant("A", "ml.p2.xlarge", accelerator_type=ACCELERATOR_TYPE), sagemaker.production_variant("B", "p299.4096xlarge", accelerator_type=ACCELERATOR_TYPE), ] ex = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" ) ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) tags = [{"ModelName": "TestModel"}] sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags ) def test_endpoint_from_production_variants_with_serverless_inference_config(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) pvs = [ sagemaker.production_variant( "A", "ml.p2.xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG ), sagemaker.production_variant( "B", "p299.4096xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG ), ] ex = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" ) ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) tags = [{"ModelName": "TestModel"}] sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags) sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags ) def test_endpoint_from_production_variants_with_async_config(sagemaker_session): ims = sagemaker_session ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"}) pvs = [ sagemaker.production_variant("A", "ml.p2.xlarge"), sagemaker.production_variant("B", "p299.4096xlarge"), ] ex = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b" ) ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex) sagemaker_session.endpoint_from_production_variants( "some-endpoint", pvs, async_inference_config_dict=AsyncInferenceConfig, ) sagemaker_session.sagemaker_client.create_endpoint.assert_called_with( EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[] ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName="some-endpoint", ProductionVariants=pvs, AsyncInferenceConfig=AsyncInferenceConfig, ) def test_update_endpoint_succeed(sagemaker_session): sagemaker_session.sagemaker_client.describe_endpoint = Mock( return_value={"EndpointStatus": "InService"} ) endpoint_name = "some-endpoint" endpoint_config = "some-endpoint-config" returned_endpoint_name = sagemaker_session.update_endpoint(endpoint_name, endpoint_config) assert returned_endpoint_name == endpoint_name def test_update_endpoint_no_wait(sagemaker_session): sagemaker_session.sagemaker_client.describe_endpoint = Mock( return_value={"EndpointStatus": "Updating"} ) endpoint_name = "some-endpoint" endpoint_config = "some-endpoint-config" returned_endpoint_name = sagemaker_session.update_endpoint( endpoint_name, endpoint_config, wait=False ) assert returned_endpoint_name == endpoint_name def test_update_endpoint_non_existing_endpoint(sagemaker_session): error = ClientError( {"Error": {"Code": "ValidationException", "Message": "Could not find entity"}}, "foo" ) expected_error_message = ( "Endpoint with name 'non-existing-endpoint' does not exist; " "please use an existing endpoint name" ) sagemaker_session.sagemaker_client.describe_endpoint = Mock(side_effect=error) with pytest.raises(ValueError, match=expected_error_message): sagemaker_session.update_endpoint("non-existing-endpoint", "non-existing-config") def test_create_endpoint_config_from_existing(sagemaker_session): pvs = [sagemaker.production_variant("A", "ml.m4.xlarge")] tags = [{"Key": "aws:cloudformation:stackname", "Value": "this-tag-should-be-ignored"}] existing_endpoint_arn = "arn:aws:sagemaker:us-west-2:123412341234:endpoint-config/foo" kms_key = "kms" sagemaker_session.sagemaker_client.describe_endpoint_config.return_value = { "Tags": tags, "ProductionVariants": pvs, "EndpointConfigArn": existing_endpoint_arn, "KmsKeyId": kms_key, } sagemaker_session.sagemaker_client.list_tags.return_value = {"Tags": tags} existing_endpoint_name = "foo" new_endpoint_name = "new-foo" sagemaker_session.create_endpoint_config_from_existing( existing_endpoint_name, new_endpoint_name ) sagemaker_session.sagemaker_client.describe_endpoint_config.assert_called_with( EndpointConfigName=existing_endpoint_name ) sagemaker_session.sagemaker_client.list_tags.assert_called_with( ResourceArn=existing_endpoint_arn, MaxResults=50 ) sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with( EndpointConfigName=new_endpoint_name, ProductionVariants=pvs, KmsKeyId=kms_key ) @patch("time.sleep") def test_wait_for_tuning_job(sleep, sagemaker_session): hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc ) result = sagemaker_session.wait_for_tuning_job(JOB_NAME) assert result["HyperParameterTuningJobStatus"] == "Completed" def test_tune_job_status(sagemaker_session): hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc ) result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME) assert result["HyperParameterTuningJobStatus"] == "Completed" def test_tune_job_status_none(sagemaker_session): hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "InProgress"} sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock( name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc ) result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME) assert result is None @patch("time.sleep") def test_wait_for_transform_job_completed(sleep, sagemaker_session): transform_job_desc = {"TransformJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_transform_job = Mock( name="describe_transform_job", return_value=transform_job_desc ) assert sagemaker_session.wait_for_transform_job(JOB_NAME)["TransformJobStatus"] == "Completed" @patch("time.sleep") def test_wait_for_transform_job_in_progress(sleep, sagemaker_session): transform_job_desc_in_progress = {"TransformJobStatus": "InProgress"} transform_job_desc_in_completed = {"TransformJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_transform_job = Mock( name="describe_transform_job", side_effect=[transform_job_desc_in_progress, transform_job_desc_in_completed], ) assert ( sagemaker_session.wait_for_transform_job(JOB_NAME, 1)["TransformJobStatus"] == "Completed" ) assert 2 == sagemaker_session.sagemaker_client.describe_transform_job.call_count def test_transform_job_status(sagemaker_session): transform_job_desc = {"TransformJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_transform_job = Mock( name="describe_transform_job", return_value=transform_job_desc ) result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME) assert result["TransformJobStatus"] == "Completed" def test_transform_job_status_none(sagemaker_session): transform_job_desc = {"TransformJobStatus": "InProgress"} sagemaker_session.sagemaker_client.describe_transform_job = Mock( name="describe_transform_job", return_value=transform_job_desc ) result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME) assert result is None def test_train_done_completed(sagemaker_session): training_job_desc = {"TrainingJobStatus": "Completed"} sagemaker_session.sagemaker_client.describe_training_job = Mock( name="describe_training_job", return_value=training_job_desc ) actual_job_desc, training_finished = _train_done( sagemaker_session.sagemaker_client, JOB_NAME, None ) assert actual_job_desc["TrainingJobStatus"] == "Completed" assert training_finished is True def test_train_done_in_progress(sagemaker_session): training_job_desc = {"TrainingJobStatus": "InProgress"} sagemaker_session.sagemaker_client.describe_training_job = Mock( name="describe_training_job", return_value=training_job_desc ) actual_job_desc, training_finished = _train_done( sagemaker_session.sagemaker_client, JOB_NAME, None ) assert actual_job_desc["TrainingJobStatus"] == "InProgress" assert training_finished is False DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS = { "AutoMLJobName": JOB_NAME, "InputDataConfig": [ { "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, "TargetAttributeName": "y", } ], "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, "AutoMLJobConfig": { "CompletionCriteria": { "MaxCandidates": 10, "MaxAutoMLJobRuntimeInSeconds": 36000, "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, } }, "RoleArn": EXPANDED_ROLE, "GenerateCandidateDefinitionsOnly": False, } COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS = { "AutoMLJobName": JOB_NAME, "InputDataConfig": [ { "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, "CompressionType": "Gzip", "TargetAttributeName": "y", } ], "OutputDataConfig": {"S3OutputPath": S3_OUTPUT}, "ProblemType": "Regression", "AutoMLJobObjective": {"Type": "type", "MetricName": "metric-name"}, "AutoMLJobConfig": { "CompletionCriteria": { "MaxCandidates": 10, "MaxAutoMLJobRuntimeInSeconds": 36000, "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, }, "SecurityConfig": { "VolumeKmsKeyId": "volume-kms-key-id-string", "EnableInterContainerTrafficEncryption": False, "VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]}, }, }, "RoleArn": EXPANDED_ROLE, "GenerateCandidateDefinitionsOnly": True, "Tags": ["tag"], } COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS = { "AutoMLJobName": JOB_NAME, "StatusEquals": "Completed", "SortOrder": "Descending", "SortBy": "Status", "MaxResults": 10, } def test_auto_ml_pack_to_request(sagemaker_session): input_config = [ { "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, "TargetAttributeName": "y", } ] output_config = {"S3OutputPath": S3_OUTPUT} auto_ml_job_config = { "CompletionCriteria": { "MaxCandidates": 10, "MaxAutoMLJobRuntimeInSeconds": 36000, "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, } } job_name = JOB_NAME role = EXPANDED_ROLE sagemaker_session.auto_ml(input_config, output_config, auto_ml_job_config, role, job_name) assert sagemaker_session.sagemaker_client.method_calls[0] == ( "create_auto_ml_job", (), DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS, ) def test_auto_ml_pack_to_request_with_optional_args(sagemaker_session): input_config = [ { "DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}}, "CompressionType": "Gzip", "TargetAttributeName": "y", } ] output_config = {"S3OutputPath": S3_OUTPUT} auto_ml_job_config = { "CompletionCriteria": { "MaxCandidates": 10, "MaxAutoMLJobRuntimeInSeconds": 36000, "MaxRuntimePerTrainingJobInSeconds": 3600 * 2, }, "SecurityConfig": { "VolumeKmsKeyId": "volume-kms-key-id-string", "EnableInterContainerTrafficEncryption": False, "VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]}, }, } job_name = JOB_NAME role = EXPANDED_ROLE sagemaker_session.auto_ml( input_config, output_config, auto_ml_job_config, role, job_name, problem_type="Regression", job_objective={"Type": "type", "MetricName": "metric-name"}, generate_candidate_definitions_only=True, tags=["tag"], ) assert sagemaker_session.sagemaker_client.method_calls[0] == ( "create_auto_ml_job", (), COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS, ) def test_list_candidates_for_auto_ml_job_default(sagemaker_session): sagemaker_session.list_candidates(job_name=JOB_NAME) sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once() sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with( AutoMLJobName=JOB_NAME ) def test_list_candidates_for_auto_ml_job_with_optional_args(sagemaker_session): sagemaker_session.list_candidates( job_name=JOB_NAME, status_equals="Completed", sort_order="Descending", sort_by="Status", max_results=10, ) sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once() sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with( **COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS ) def test_describe_tuning_job(sagemaker_session): job_name = "hyper-parameter-tuning" sagemaker_session.describe_tuning_job(job_name=job_name) sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job.assert_called_with( HyperParameterTuningJobName=job_name ) @pytest.fixture def pipeline_empty_definition(): return '{"Version": "2020-12-01", ' '"Metadata": {}, ' '"Parameters": [], ' '"Steps": []}' @pytest.fixture def pipeline_role_arn(): return "my:pipeline:role:arn" def test_describe_model(sagemaker_session): model_name = "sagemaker-model-name" sagemaker_session.describe_model(name=model_name) sagemaker_session.sagemaker_client.describe_model.assert_called_with(ModelName=model_name) def test_create_model_package_from_containers(sagemaker_session): model_package_name = "sagemaker-model-package" sagemaker_session.create_model_package_from_containers(model_package_name=model_package_name) sagemaker_session.sagemaker_client.create_model_package.assert_called_once() def test_create_model_package_from_containers_name_conflict(sagemaker_session): model_package_name = "sagemaker-model-package" model_package_group_name = "sagemaker-model-package-group" with pytest.raises(ValueError) as error: sagemaker_session.create_model_package_from_containers( model_package_name=model_package_name, model_package_group_name=model_package_group_name, ) assert ( "model_package_name and model_package_group_name cannot be present at the same " "time." == str(error) ) def test_create_model_package_from_containers_incomplete_args(sagemaker_session): model_package_name = "sagemaker-model-package" containers = ["dummy-container"] with pytest.raises(ValueError) as error: sagemaker_session.create_model_package_from_containers( model_package_name=model_package_name, containers=containers, ) assert ( "content_types and response_types " "must be provided if containers is present." == str(error) ) def test_create_model_package_from_containers_without_model_package_group_name(sagemaker_session): model_package_name = "sagemaker-model-package" containers = ["dummy-container"] content_types = ["application/json"] response_types = ["application/json"] with pytest.raises(ValueError) as error: sagemaker_session.create_model_package_from_containers( model_package_name=model_package_name, containers=containers, content_types=content_types, response_types=response_types, ) assert ( "inference_inferences and transform_instances " "must be provided if model_package_group_name is not present." == str(error) ) def test_create_model_package_from_containers_all_args(sagemaker_session): model_package_name = "sagemaker-model-package" containers = ["dummy-container"] content_types = ["application/json"] response_types = ["application/json"] inference_instances = ["ml.m4.xlarge"] transform_instances = ["ml.m4.xlarget"] model_metrics = { "Bias": { "ContentType": "content-type", "S3Uri": "s3://...", } } drift_check_baselines = { "Bias": { "ConfigFile": { "ContentType": "content-type", "S3Uri": "s3://...", } } } metadata_properties = { "CommitId": "test-commit-id", "Repository": "test-repository", "GeneratedBy": "sagemaker-python-sdk", "ProjectId": "unit-test", } marketplace_cert = (True,) approval_status = ("Approved",) description = "description" customer_metadata_properties = {"key1": "value1"} domain = "COMPUTER_VISION" task = "IMAGE_CLASSIFICATION" sample_payload_url = "s3://test-bucket/model" sagemaker_session.create_model_package_from_containers( containers=containers, content_types=content_types, response_types=response_types, inference_instances=inference_instances, transform_instances=transform_instances, model_package_name=model_package_name, model_metrics=model_metrics, metadata_properties=metadata_properties, marketplace_cert=marketplace_cert, approval_status=approval_status, description=description, drift_check_baselines=drift_check_baselines, customer_metadata_properties=customer_metadata_properties, domain=domain, sample_payload_url=sample_payload_url, task=task, ) expected_args = { "ModelPackageName": model_package_name, "InferenceSpecification": { "Containers": containers, "SupportedContentTypes": content_types, "SupportedResponseMIMETypes": response_types, "SupportedRealtimeInferenceInstanceTypes": inference_instances, "SupportedTransformInstanceTypes": transform_instances, }, "ModelPackageDescription": description, "ModelMetrics": model_metrics, "MetadataProperties": metadata_properties, "CertifyForMarketplace": marketplace_cert, "ModelApprovalStatus": approval_status, "DriftCheckBaselines": drift_check_baselines, "CustomerMetadataProperties": customer_metadata_properties, "Domain": domain, "SamplePayloadUrl": sample_payload_url, "Task": task, } sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) def test_create_model_package_from_containers_without_instance_types(sagemaker_session): model_package_group_name = "sagemaker-model-package-group-name-1.0" containers = ["dummy-container"] content_types = ["application/json"] response_types = ["application/json"] model_metrics = { "Bias": { "ContentType": "content-type", "S3Uri": "s3://...", } } drift_check_baselines = { "Bias": { "ConfigFile": { "ContentType": "content-type", "S3Uri": "s3://...", } } } metadata_properties = { "CommitId": "test-commit-id", "Repository": "test-repository", "GeneratedBy": "sagemaker-python-sdk", "ProjectId": "unit-test", } marketplace_cert = (True,) approval_status = ("Approved",) description = "description" customer_metadata_properties = {"key1": "value1"} sagemaker_session.create_model_package_from_containers( containers=containers, content_types=content_types, response_types=response_types, model_package_group_name=model_package_group_name, model_metrics=model_metrics, metadata_properties=metadata_properties, marketplace_cert=marketplace_cert, approval_status=approval_status, description=description, drift_check_baselines=drift_check_baselines, customer_metadata_properties=customer_metadata_properties, ) expected_args = { "ModelPackageGroupName": model_package_group_name, "InferenceSpecification": { "Containers": containers, "SupportedContentTypes": content_types, "SupportedResponseMIMETypes": response_types, }, "ModelPackageDescription": description, "ModelMetrics": model_metrics, "MetadataProperties": metadata_properties, "CertifyForMarketplace": marketplace_cert, "ModelApprovalStatus": approval_status, "DriftCheckBaselines": drift_check_baselines, "CustomerMetadataProperties": customer_metadata_properties, } sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) def test_create_model_package_from_containers_with_one_instance_types(sagemaker_session): model_package_group_name = "sagemaker-model-package-group-name-1.0" containers = ["dummy-container"] content_types = ["application/json"] response_types = ["application/json"] transform_instances = ["ml.m5.xlarge"] model_metrics = { "Bias": { "ContentType": "content-type", "S3Uri": "s3://...", } } drift_check_baselines = { "Bias": { "ConfigFile": { "ContentType": "content-type", "S3Uri": "s3://...", } } } metadata_properties = { "CommitId": "test-commit-id", "Repository": "test-repository", "GeneratedBy": "sagemaker-python-sdk", "ProjectId": "unit-test", } marketplace_cert = (True,) approval_status = ("Approved",) description = "description" customer_metadata_properties = {"key1": "value1"} sagemaker_session.create_model_package_from_containers( containers=containers, content_types=content_types, response_types=response_types, transform_instances=transform_instances, model_package_group_name=model_package_group_name, model_metrics=model_metrics, metadata_properties=metadata_properties, marketplace_cert=marketplace_cert, approval_status=approval_status, description=description, drift_check_baselines=drift_check_baselines, customer_metadata_properties=customer_metadata_properties, ) expected_args = { "ModelPackageGroupName": model_package_group_name, "InferenceSpecification": { "Containers": containers, "SupportedContentTypes": content_types, "SupportedResponseMIMETypes": response_types, "SupportedTransformInstanceTypes": transform_instances, }, "ModelPackageDescription": description, "ModelMetrics": model_metrics, "MetadataProperties": metadata_properties, "CertifyForMarketplace": marketplace_cert, "ModelApprovalStatus": approval_status, "DriftCheckBaselines": drift_check_baselines, "CustomerMetadataProperties": customer_metadata_properties, } sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args) @pytest.fixture def feature_group_dummy_definitions(): return [{"FeatureName": "feature1", "FeatureType": "String"}] def test_feature_group_create(sagemaker_session, feature_group_dummy_definitions): sagemaker_session.create_feature_group( feature_group_name="MyFeatureGroup", record_identifier_name="feature1", event_time_feature_name="feature2", feature_definitions=feature_group_dummy_definitions, role_arn="dummy_role", ) assert sagemaker_session.sagemaker_client.create_feature_group.called_with( FeatureGroupName="MyFeatureGroup", RecordIdentifierFeatureName="feature1", EventTimeFeatureName="feature2", FeatureDefinitions=feature_group_dummy_definitions, RoleArn="dummy_role", ) def test_feature_group_delete(sagemaker_session): sagemaker_session.delete_feature_group(feature_group_name="MyFeatureGroup") assert sagemaker_session.sagemaker_client.delete_feature_group.called_with( FeatureGroupName="MyFeatureGroup", ) def test_feature_group_describe(sagemaker_session): sagemaker_session.describe_feature_group(feature_group_name="MyFeatureGroup") assert sagemaker_session.sagemaker_client.describe_feature_group.called_with( FeatureGroupName="MyFeatureGroup", ) def test_feature_group_update(sagemaker_session, feature_group_dummy_definitions): sagemaker_session.update_feature_group( feature_group_name="MyFeatureGroup", feature_additions=feature_group_dummy_definitions, ) assert sagemaker_session.sagemaker_client.update_feature_group.called_with( FeatureGroupName="MyFeatureGroup", FeatureAdditions=feature_group_dummy_definitions, ) def test_feature_metadata_update(sagemaker_session): parameter_additions = [ { "key": "TestKey", "value": "TestValue", } ] parameter_removals = ["TestKey"] sagemaker_session.update_feature_metadata( feature_group_name="TestFeatureGroup", feature_name="TestFeature", description="TestDescription", parameter_additions=parameter_additions, parameter_removals=parameter_removals, ) assert sagemaker_session.sagemaker_client.update_feature_group.called_with( feature_group_name="TestFeatureGroup", FeatureName="TestFeature", Description="TestDescription", ParameterAdditions=parameter_additions, ParameterRemovals=parameter_removals, ) sagemaker_session.update_feature_metadata( feature_group_name="TestFeatureGroup", feature_name="TestFeature", ) assert sagemaker_session.sagemaker_client.update_feature_group.called_with( feature_group_name="TestFeatureGroup", FeatureName="TestFeature", ) def test_feature_metadata_describe(sagemaker_session): sagemaker_session.describe_feature_metadata( feature_group_name="MyFeatureGroup", feature_name="TestFeature" ) assert sagemaker_session.sagemaker_client.describe_feature_metadata.called_with( FeatureGroupName="MyFeatureGroup", FeatureName="TestFeature" ) def test_start_query_execution(sagemaker_session): athena_mock = Mock() sagemaker_session.boto_session.client( "athena", region_name=sagemaker_session.boto_region_name ).return_value = athena_mock sagemaker_session.start_query_execution( catalog="catalog", database="database", query_string="query", output_location="s3://results", ) assert athena_mock.start_query_execution.called_once_with( QueryString="query", QueryExecutionContext={"Catalog": "catalog", "Database": "database"}, OutputLocation="s3://results", ) def test_get_query_execution(sagemaker_session): athena_mock = Mock() sagemaker_session.boto_session.client( "athena", region_name=sagemaker_session.boto_region_name ).return_value = athena_mock sagemaker_session.get_query_execution(query_execution_id="query_id") assert athena_mock.get_query_execution.called_with(QueryExecutionId="query_id") def test_download_athena_query_result(sagemaker_session): sagemaker_session.s3_client = Mock() sagemaker_session.download_athena_query_result( bucket="bucket", prefix="prefix", query_execution_id="query_id", filename="filename", ) assert sagemaker_session.s3_client.download_file.called_with( Bucket="bucket", Key="prefix/query_id.csv", Filename="filename", ) @patch("sagemaker.session.Session.get_query_execution") def test_wait_for_athena_query(query_execution, sagemaker_session): query_execution.return_value = {"QueryExecution": {"Status": {"State": "SUCCEEDED"}}} sagemaker_session.wait_for_athena_query(query_execution_id="query_id") assert query_execution.called_with(query_execution_id="query_id")