hc99's picture
Add files using upload-large-folder tool
476455e verified
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"). You
# may not use this file except in compliance with the License. A copy of
# the License is located at
#
# http://aws.amazon.com/apache2.0/
#
# or in the "license" file accompanying this file. This file is
# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
# ANY KIND, either express or implied. See the License for the specific
# language governing permissions and limitations under the License.
from __future__ import absolute_import
from copy import deepcopy
import logging
import json
import os
import subprocess
from time import sleep
from sagemaker.fw_utils import UploadedCode
import pytest
from botocore.exceptions import ClientError
from mock import ANY, MagicMock, Mock, patch, PropertyMock
from sagemaker.huggingface.estimator import HuggingFace
from sagemaker.jumpstart.constants import JUMPSTART_BUCKET_NAME_SET, JUMPSTART_RESOURCE_BASE_NAME
from sagemaker.jumpstart.enums import JumpStartTag
import sagemaker.local
from sagemaker import TrainingInput, utils, vpc_utils
from sagemaker.algorithm import AlgorithmEstimator
from sagemaker.debugger import (
rule_configs,
CollectionConfig,
DebuggerHookConfig,
FrameworkProfile,
ProfilerConfig,
ProfilerRule,
Rule,
)
from sagemaker.async_inference import AsyncInferenceConfig
from sagemaker.estimator import Estimator, EstimatorBase, Framework, _TrainingJob
from sagemaker.fw_utils import PROFILER_UNSUPPORTED_REGIONS
from sagemaker.inputs import ShuffleConfig
from sagemaker.instance_group import InstanceGroup
from sagemaker.model import FrameworkModel
from sagemaker.mxnet.estimator import MXNet
from sagemaker.predictor import Predictor
from sagemaker.pytorch.estimator import PyTorch
from sagemaker.sklearn.estimator import SKLearn
from sagemaker.tensorflow.estimator import TensorFlow
from sagemaker.predictor_async import AsyncPredictor
from sagemaker.transformer import Transformer
from sagemaker.workflow.parameters import ParameterString, ParameterBoolean
from sagemaker.workflow.pipeline_context import PipelineSession
from sagemaker.xgboost.estimator import XGBoost
MODEL_DATA = "s3://bucket/model.tar.gz"
MODEL_IMAGE = "mi"
ENTRY_POINT = "blah.py"
DATA_DIR = os.path.join(os.path.dirname(__file__), "..", "data")
SCRIPT_NAME = "dummy_script.py"
SCRIPT_PATH = os.path.join(DATA_DIR, SCRIPT_NAME)
TIMESTAMP = "2017-11-06-14:14:15.671"
TIME = 1510006209.073025
BUCKET_NAME = "mybucket"
INSTANCE_COUNT = 1
INSTANCE_TYPE = "c4.4xlarge"
KEEP_ALIVE_PERIOD_IN_SECONDS = 1800
ACCELERATOR_TYPE = "ml.eia.medium"
ROLE = "DummyRole"
IMAGE_URI = "fakeimage"
REGION = "us-west-2"
JOB_NAME = "{}-{}".format(IMAGE_URI, TIMESTAMP)
TAGS = [{"Name": "some-tag", "Value": "value-for-tag"}]
OUTPUT_PATH = "s3://bucket/prefix"
GIT_REPO = "https://github.com/aws/sagemaker-python-sdk.git"
BRANCH = "test-branch-git-config"
COMMIT = "ae15c9d7d5b97ea95ea451e4662ee43da3401d73"
PRIVATE_GIT_REPO_SSH = "git@github.com:testAccount/private-repo.git"
PRIVATE_GIT_REPO = "https://github.com/testAccount/private-repo.git"
PRIVATE_BRANCH = "test-branch"
PRIVATE_COMMIT = "329bfcf884482002c05ff7f44f62599ebc9f445a"
CODECOMMIT_REPO = "https://git-codecommit.us-west-2.amazonaws.com/v1/repos/test-repo/"
CODECOMMIT_REPO_SSH = "ssh://git-codecommit.us-west-2.amazonaws.com/v1/repos/test-repo/"
CODECOMMIT_BRANCH = "master"
REPO_DIR = "/tmp/repo_dir"
ENV_INPUT = {"env_key1": "env_val1", "env_key2": "env_val2", "env_key3": "env_val3"}
DESCRIBE_TRAINING_JOB_RESULT = {"ModelArtifacts": {"S3ModelArtifacts": MODEL_DATA}}
RETURNED_JOB_DESCRIPTION = {
"AlgorithmSpecification": {
"TrainingInputMode": "File",
"TrainingImage": "1.dkr.ecr.us-west-2.amazonaws.com/sagemaker-other:1.0.4",
},
"HyperParameters": {
"sagemaker_submit_directory": '"s3://some/sourcedir.tar.gz"',
"checkpoint_path": '"s3://other/1508872349"',
"sagemaker_program": '"iris-dnn-classifier.py"',
"sagemaker_container_log_level": '"logging.INFO"',
"sagemaker_job_name": '"neo"',
"training_steps": "100",
},
"RoleArn": "arn:aws:iam::366:role/SageMakerRole",
"ResourceConfig": {"VolumeSizeInGB": 30, "InstanceCount": 1, "InstanceType": "ml.c4.xlarge"},
"EnableNetworkIsolation": False,
"StoppingCondition": {"MaxRuntimeInSeconds": 24 * 60 * 60},
"TrainingJobName": "neo",
"TrainingJobStatus": "Completed",
"TrainingJobArn": "arn:aws:sagemaker:us-west-2:336:training-job/neo",
"OutputDataConfig": {"KmsKeyId": "", "S3OutputPath": "s3://place/output/neo"},
"TrainingJobOutput": {"S3TrainingJobOutput": "s3://here/output.tar.gz"},
"EnableInterContainerTrafficEncryption": False,
}
MODEL_CONTAINER_DEF = {
"Environment": {
"SAGEMAKER_PROGRAM": ENTRY_POINT,
"SAGEMAKER_SUBMIT_DIRECTORY": "s3://mybucket/mi-2017-10-10-14-14-15/sourcedir.tar.gz",
"SAGEMAKER_CONTAINER_LOG_LEVEL": "20",
"SAGEMAKER_REGION": REGION,
},
"Image": MODEL_IMAGE,
"ModelDataUrl": MODEL_DATA,
}
ENDPOINT_DESC = {"EndpointConfigName": "test-endpoint"}
ENDPOINT_CONFIG_DESC = {"ProductionVariants": [{"ModelName": "model-1"}, {"ModelName": "model-2"}]}
LIST_TAGS_RESULT = {"Tags": [{"Key": "TagtestKey", "Value": "TagtestValue"}]}
DISTRIBUTION_PS_ENABLED = {"parameter_server": {"enabled": True}}
DISTRIBUTION_MPI_ENABLED = {
"mpi": {"enabled": True, "custom_mpi_options": "options", "processes_per_host": 2}
}
DISTRIBUTION_SM_DDP_ENABLED = {
"smdistributed": {"dataparallel": {"enabled": True, "custom_mpi_options": "options"}}
}
class DummyFramework(Framework):
_framework_name = "dummy"
def training_image_uri(self):
return IMAGE_URI
def create_model(
self,
role=None,
model_server_workers=None,
entry_point=None,
vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT,
enable_network_isolation=None,
model_dir=None,
**kwargs,
):
if enable_network_isolation is None:
enable_network_isolation = self.enable_network_isolation()
return DummyFrameworkModel(
self.sagemaker_session,
vpc_config=self.get_vpc_config(vpc_config_override),
entry_point=entry_point,
enable_network_isolation=enable_network_isolation,
role=role,
**kwargs,
)
@classmethod
def _prepare_init_params_from_job_description(cls, job_details, model_channel_name=None):
init_params = super(DummyFramework, cls)._prepare_init_params_from_job_description(
job_details, model_channel_name
)
init_params.pop("image_uri", None)
return init_params
class DummyFrameworkModel(FrameworkModel):
def __init__(self, sagemaker_session, entry_point=None, role=ROLE, **kwargs):
super(DummyFrameworkModel, self).__init__(
MODEL_DATA,
MODEL_IMAGE,
role,
entry_point or ENTRY_POINT,
sagemaker_session=sagemaker_session,
**kwargs,
)
def create_predictor(self, endpoint_name):
return None
def prepare_container_def(
self, instance_type, accelerator_type=None, serverless_inference_config=None
):
return MODEL_CONTAINER_DEF
@pytest.fixture(autouse=True)
def mock_create_tar_file():
with patch("sagemaker.utils.create_tar_file", MagicMock()) as create_tar_file:
yield create_tar_file
@pytest.fixture()
def sagemaker_session():
boto_mock = Mock(name="boto_session", region_name=REGION)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=REGION,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
sms.default_bucket = Mock(name="default_bucket", return_value=BUCKET_NAME)
sms.sagemaker_client.describe_training_job = Mock(
name="describe_training_job", return_value=DESCRIBE_TRAINING_JOB_RESULT
)
sms.sagemaker_client.describe_endpoint = Mock(return_value=ENDPOINT_DESC)
sms.sagemaker_client.describe_endpoint_config = Mock(return_value=ENDPOINT_CONFIG_DESC)
sms.sagemaker_client.list_tags = Mock(return_value=LIST_TAGS_RESULT)
sms.upload_data = Mock(return_value=OUTPUT_PATH)
return sms
@pytest.fixture()
def pipeline_session():
client_mock = Mock()
client_mock._client_config.user_agent = (
"Boto3/1.14.24 Python/3.8.5 Linux/5.4.0-42-generic Botocore/1.17.24 Resource"
)
role_mock = Mock()
type(role_mock).arn = PropertyMock(return_value=ROLE)
resource_mock = Mock()
resource_mock.Role.return_value = role_mock
session_mock = Mock(region_name=REGION)
session_mock.resource.return_value = resource_mock
session_mock.client.return_value = client_mock
return PipelineSession(
boto_session=session_mock, sagemaker_client=client_mock, default_bucket=BUCKET_NAME
)
@pytest.fixture()
def training_job_description(sagemaker_session):
returned_job_description = RETURNED_JOB_DESCRIPTION.copy()
mock_describe_training_job = Mock(
name="describe_training_job", return_value=returned_job_description
)
sagemaker_session.sagemaker_client.describe_training_job = mock_describe_training_job
sagemaker_session.describe_training_job = mock_describe_training_job
return returned_job_description
def test_framework_all_init_args(sagemaker_session):
f = DummyFramework(
"my_script.py",
role="DummyRole",
instance_count=3,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
volume_size=123,
volume_kms_key="volumekms",
max_run=456,
input_mode="inputmode",
output_path="outputpath",
output_kms_key="outputkms",
base_job_name="basejobname",
tags=[{"foo": "bar"}],
subnets=["123", "456"],
security_group_ids=["789", "012"],
metric_definitions=[{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}],
encrypt_inter_container_traffic=True,
checkpoint_s3_uri="s3://bucket/checkpoint",
checkpoint_local_path="file://local/checkpoint",
enable_sagemaker_metrics=True,
enable_network_isolation=True,
environment=ENV_INPUT,
max_retry_attempts=2,
)
_TrainingJob.start_new(f, "s3://mydata", None)
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args == {
"input_mode": "inputmode",
"tags": [{"foo": "bar"}],
"hyperparameters": {},
"image_uri": "fakeimage",
"input_config": [
{
"ChannelName": "training",
"DataSource": {
"S3DataSource": {
"S3DataType": "S3Prefix",
"S3DataDistributionType": "FullyReplicated",
"S3Uri": "s3://mydata",
}
},
}
],
"output_config": {"KmsKeyId": "outputkms", "S3OutputPath": "outputpath"},
"vpc_config": {"Subnets": ["123", "456"], "SecurityGroupIds": ["789", "012"]},
"stop_condition": {"MaxRuntimeInSeconds": 456},
"retry_strategy": {"MaximumRetryAttempts": 2},
"role": sagemaker_session.expand_role(),
"job_name": None,
"resource_config": {
"VolumeSizeInGB": 123,
"InstanceCount": 3,
"VolumeKmsKeyId": "volumekms",
"InstanceType": "ml.m4.xlarge",
},
"metric_definitions": [{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}],
"encrypt_inter_container_traffic": True,
"environment": {"env_key1": "env_val1", "env_key2": "env_val2", "env_key3": "env_val3"},
"experiment_config": None,
"checkpoint_s3_uri": "s3://bucket/checkpoint",
"checkpoint_local_path": "file://local/checkpoint",
"enable_sagemaker_metrics": True,
"enable_network_isolation": True,
}
def test_framework_with_heterogeneous_cluster(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_groups=[
InstanceGroup("group1", "ml.c4.xlarge", 1),
InstanceGroup("group2", "ml.m4.xlarge", 2),
],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["resource_config"]["InstanceGroups"][0] == {
"InstanceGroupName": "group1",
"InstanceCount": 1,
"InstanceType": "ml.c4.xlarge",
}
assert args["resource_config"]["InstanceGroups"][1] == {
"InstanceGroupName": "group2",
"InstanceCount": 2,
"InstanceType": "ml.m4.xlarge",
}
def test_framework_with_keep_alive_period(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_groups=[
InstanceGroup("group1", "ml.c4.xlarge", 1),
InstanceGroup("group2", "ml.m4.xlarge", 2),
],
keep_alive_period_in_seconds=KEEP_ALIVE_PERIOD_IN_SECONDS,
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["resource_config"]["KeepAlivePeriodInSeconds"] == KEEP_ALIVE_PERIOD_IN_SECONDS
def test_framework_with_debugger_and_built_in_rule(sagemaker_session):
debugger_built_in_rule_with_custom_args = Rule.sagemaker(
base_config=rule_configs.stalled_training_rule(),
rule_parameters={"threshold": "120", "stop_training_on_fire": "True"},
collections_to_save=[
CollectionConfig(
name="losses", parameters={"train.save_interval": "50", "eval.save_interval": "10"}
)
],
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[debugger_built_in_rule_with_custom_args],
debugger_hook_config=DebuggerHookConfig(s3_output_path="s3://output"),
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"][0]["RuleParameters"] == {
"rule_to_invoke": "StalledTrainingRule",
"threshold": "120",
"stop_training_on_fire": "True",
}
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://output",
"CollectionConfigurations": [
{
"CollectionName": "losses",
"CollectionParameters": {"train.save_interval": "50", "eval.save_interval": "10"},
}
],
}
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
def test_framework_with_debugger_and_custom_rule(sagemaker_session):
hook_config = DebuggerHookConfig(
s3_output_path="s3://output", collection_configs=[CollectionConfig(name="weights")]
)
debugger_custom_rule = Rule.custom(
name="CustomRule",
image_uri="RuleImageUri",
instance_type=INSTANCE_TYPE,
volume_size_in_gb=5,
source="path/to/my_custom_rule.py",
rule_to_invoke="CustomRule",
other_trials_s3_input_paths=["s3://path/trial1", "s3://path/trial2"],
rule_parameters={"threshold": "120"},
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[debugger_custom_rule],
debugger_hook_config=hook_config,
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"] == [
{
"RuleConfigurationName": "CustomRule",
"RuleEvaluatorImage": "RuleImageUri",
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": 5,
"RuleParameters": {
"source_s3_uri": sagemaker_session.upload_data(),
"rule_to_invoke": "CustomRule",
"threshold": "120",
"other_trial_0": "s3://path/trial1",
"other_trial_1": "s3://path/trial2",
},
}
]
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://output",
"CollectionConfigurations": [{"CollectionName": "weights"}],
}
def test_framework_with_only_debugger_rule(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[Rule.sagemaker(rule_configs.stalled_training_rule())],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"][0]["RuleParameters"] == {
"rule_to_invoke": "StalledTrainingRule"
}
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
"CollectionConfigurations": [],
}
def test_framework_with_debugger_rule_and_single_action(sagemaker_session):
stop_training_action = rule_configs.StopTraining()
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[Rule.sagemaker(rule_configs.stalled_training_rule(), actions=stop_training_action)],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"][0]["RuleParameters"] == {
"rule_to_invoke": "StalledTrainingRule",
"action_json": stop_training_action.serialize(),
}
assert stop_training_action.action_parameters["training_job_prefix"] == f._current_job_name
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
"CollectionConfigurations": [],
}
def test_framework_with_debugger_rule_and_multiple_actions(sagemaker_session):
action_list = rule_configs.ActionList(
rule_configs.StopTraining(),
rule_configs.Email("abc@abc.com"),
rule_configs.SMS("+1234567890"),
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[Rule.sagemaker(rule_configs.stalled_training_rule(), actions=action_list)],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"][0]["RuleParameters"] == {
"rule_to_invoke": "StalledTrainingRule",
"action_json": action_list.serialize(),
}
assert action_list.actions[0].action_parameters["training_job_prefix"] == f._current_job_name
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
"CollectionConfigurations": [],
}
def test_framework_with_only_debugger_hook_config(sagemaker_session):
hook_config = DebuggerHookConfig(
s3_output_path="s3://output", collection_configs=[CollectionConfig(name="weights")]
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
debugger_hook_config=hook_config,
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://output",
"CollectionConfigurations": [{"CollectionName": "weights"}],
}
assert "debugger_rule_configs" not in args
@patch("time.time", return_value=TIME)
def test_framework_without_debugger_and_profiler(time, sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_hook_config"] == {
"CollectionConfigurations": [],
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
assert "debugger_rule_configs" not in args
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "ProfilerReport-1510006209",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
]
def test_framework_with_debugger_and_profiler_rules(sagemaker_session):
debugger_built_in_rule_with_custom_args = Rule.sagemaker(
base_config=rule_configs.stalled_training_rule(),
rule_parameters={"threshold": "120", "stop_training_on_fire": "True"},
collections_to_save=[
CollectionConfig(
name="losses", parameters={"train.save_interval": "50", "eval.save_interval": "10"}
)
],
)
profiler_built_in_rule_with_custom_args = ProfilerRule.sagemaker(
base_config=rule_configs.ProfilerReport(CPUBottleneck_threshold=90),
name="CustomProfilerReportRule",
)
profiler_custom_rule = ProfilerRule.custom(
name="CustomProfilerRule",
image_uri="RuleImageUri",
instance_type=INSTANCE_TYPE,
volume_size_in_gb=5,
source="path/to/my_custom_rule.py",
rule_to_invoke="CustomProfilerRule",
rule_parameters={"threshold": "10"},
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[
debugger_built_in_rule_with_custom_args,
profiler_built_in_rule_with_custom_args,
profiler_custom_rule,
],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["debugger_rule_configs"] == [
{
"RuleConfigurationName": "StalledTrainingRule",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {
"rule_to_invoke": "StalledTrainingRule",
"threshold": "120",
"stop_training_on_fire": "True",
},
}
]
assert args["debugger_hook_config"] == {
"S3OutputPath": "s3://mybucket/",
"CollectionConfigurations": [
{
"CollectionName": "losses",
"CollectionParameters": {"train.save_interval": "50", "eval.save_interval": "10"},
}
],
}
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "CustomProfilerReportRule",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport", "CPUBottleneck_threshold": "90"},
},
{
"InstanceType": "c4.4xlarge",
"RuleConfigurationName": "CustomProfilerRule",
"RuleEvaluatorImage": "RuleImageUri",
"RuleParameters": {
"rule_to_invoke": "CustomProfilerRule",
"source_s3_uri": OUTPUT_PATH,
"threshold": "10",
},
"VolumeSizeInGB": 5,
},
]
def test_framework_with_only_profiler_rule_specified(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[ProfilerRule.sagemaker(rule_configs.CPUBottleneck(gpu_threshold=60))],
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "CPUBottleneck",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {
"rule_to_invoke": "CPUBottleneck",
"cpu_threshold": "90",
"gpu_threshold": "60",
"patience": "1000",
"scan_interval_us": "60000000",
"threshold": "50",
},
}
]
@patch("time.time", return_value=TIME)
def test_framework_with_profiler_config_without_s3_output_path(time, sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
profiler_config=ProfilerConfig(system_monitor_interval_millis=1000),
)
f.fit("s3://mydata")
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
"ProfilingIntervalInMilliseconds": 1000,
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "ProfilerReport-1510006209",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
]
@pytest.mark.parametrize("region", PROFILER_UNSUPPORTED_REGIONS)
def test_framework_with_no_default_profiler_in_unsupported_region(region):
boto_mock = Mock(name="boto_session", region_name=region)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=region,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
sms.train.assert_called_once()
_, args = sms.train.call_args
assert args.get("profiler_config") is None
assert args.get("profiler_rule_configs") is None
@pytest.mark.parametrize("region", PROFILER_UNSUPPORTED_REGIONS)
def test_framework_with_debugger_config_set_up_in_unsupported_region(region):
with pytest.raises(ValueError) as error:
boto_mock = Mock(name="boto_session", region_name=region)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=region,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
debugger_hook_config=DebuggerHookConfig(s3_output_path="s3://output"),
)
f.fit("s3://mydata")
assert "Current region does not support debugger but debugger hook config is set!" in str(error)
@pytest.mark.parametrize("region", PROFILER_UNSUPPORTED_REGIONS)
def test_framework_enable_profiling_in_unsupported_region(region):
with pytest.raises(ValueError) as error:
boto_mock = Mock(name="boto_session", region_name=region)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=region,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.enable_default_profiling()
assert "Current region does not support profiler / debugger!" in str(error)
@pytest.mark.parametrize("region", PROFILER_UNSUPPORTED_REGIONS)
def test_framework_update_profiling_in_unsupported_region(region):
with pytest.raises(ValueError) as error:
boto_mock = Mock(name="boto_session", region_name=region)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=region,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(system_monitor_interval_millis=1000)
assert "Current region does not support profiler / debugger!" in str(error)
@pytest.mark.parametrize("region", PROFILER_UNSUPPORTED_REGIONS)
def test_framework_disable_profiling_in_unsupported_region(region):
with pytest.raises(ValueError) as error:
boto_mock = Mock(name="boto_session", region_name=region)
sms = MagicMock(
name="sagemaker_session",
boto_session=boto_mock,
boto_region_name=region,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.disable_profiling()
assert "Current region does not support profiler / debugger!" in str(error)
def test_framework_with_profiler_config_and_profiler_disabled(sagemaker_session):
with pytest.raises(RuntimeError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
profiler_config=ProfilerConfig(),
disable_profiler=True,
)
f.fit("s3://mydata")
assert "profiler_config cannot be set when disable_profiler is True." in str(error)
def test_framework_with_profiler_rule_and_profiler_disabled(sagemaker_session):
profiler_custom_rule = ProfilerRule.custom(
name="CustomProfilerRule",
image_uri="RuleImageUri",
instance_type=INSTANCE_TYPE,
volume_size_in_gb=5,
)
with pytest.raises(RuntimeError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
rules=[profiler_custom_rule],
disable_profiler=True,
)
f.fit("s3://mydata")
assert "ProfilerRule cannot be set when disable_profiler is True." in str(error)
def test_framework_with_enabling_default_profiling_when_profiler_is_already_enabled(
sagemaker_session, training_job_description
):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
training_job_description["ProfilingStatus"] = "Enabled"
f.enable_default_profiling()
assert (
"Debugger monitoring is already enabled. To update the profiler_config parameter "
"and the Debugger profiling rules, please use the update_profiler function." in str(error)
)
@patch("time.time", return_value=TIME)
def test_framework_with_enabling_default_profiling(
time, sagemaker_session, training_job_description
):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
disable_profiler=True,
)
f.fit("s3://mydata")
training_job_description["ProfilingStatus"] = "Disabled"
f.enable_default_profiling()
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {
"S3OutputPath": "s3://{}/".format(BUCKET_NAME),
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "ProfilerReport-1510006209",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
]
@patch("time.time", return_value=TIME)
def test_framework_with_enabling_default_profiling_with_existed_s3_output_path(
time, sagemaker_session, training_job_description
):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
disable_profiler=True,
)
f.fit("s3://mydata")
training_job_description["ProfilingStatus"] = "Disabled"
training_job_description["ProfilerConfig"] = {
"S3OutputPath": "s3://custom/",
"ProfilingIntervalInMilliseconds": 1000,
}
f.enable_default_profiling()
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {
"S3OutputPath": "s3://custom/",
}
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "ProfilerReport-1510006209",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
]
def test_framework_with_disabling_profiling_when_profiler_is_already_disabled(
sagemaker_session, training_job_description
):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
training_job_description["ProfilingStatus"] = "Disabled"
f.disable_profiling()
assert "Profiler is already disabled." in str(error)
def test_framework_with_disabling_profiling(sagemaker_session, training_job_description):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
training_job_description["ProfilingStatus"] = "Enabled"
f.disable_profiling()
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {"DisableProfiler": True}
def test_framework_with_update_profiler_when_no_training_job(sagemaker_session):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.update_profiler(system_monitor_interval_millis=1000)
assert "Estimator is not associated with a training job" in str(error)
def test_framework_with_update_profiler_without_any_parameter(sagemaker_session):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler()
assert "Please provide profiler config or profiler rule to be updated." in str(error)
def test_framework_with_update_profiler_with_debugger_rule(sagemaker_session):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(rules=[Rule.sagemaker(rule_configs.stalled_training_rule())])
assert "Please provide ProfilerRule to be updated." in str(error)
def test_framework_with_update_profiler_config(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(system_monitor_interval_millis=1000)
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {
"ProfilingIntervalInMilliseconds": 1000,
}
assert "profiler_rule_configs" not in args
def test_framework_with_update_profiler_report_rule(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(
rules=[
ProfilerRule.sagemaker(rule_configs.ProfilerReport(), name="CustomProfilerReportRule")
]
)
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_rule_configs"] == [
{
"RuleConfigurationName": "CustomProfilerReportRule",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
]
assert "profiler_config" not in args
def test_framework_with_disable_framework_metrics(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(disable_framework_metrics=True)
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {"ProfilingParameters": {}}
assert "profiler_rule_configs" not in args
def test_framework_with_disable_framework_metrics_and_update_system_metrics(sagemaker_session):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(system_monitor_interval_millis=1000, disable_framework_metrics=True)
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {
"ProfilingIntervalInMilliseconds": 1000,
"ProfilingParameters": {},
}
assert "profiler_rule_configs" not in args
def test_framework_with_disable_framework_metrics_and_update_framework_params(sagemaker_session):
with pytest.raises(ValueError) as error:
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(
framework_profile_params=FrameworkProfile(), disable_framework_metrics=True
)
assert "framework_profile_params cannot be set when disable_framework_metrics is True" in str(
error
)
def test_framework_with_update_profiler_config_and_profiler_rule(sagemaker_session):
profiler_custom_rule = ProfilerRule.custom(
name="CustomProfilerRule",
image_uri="RuleImageUri",
instance_type=INSTANCE_TYPE,
volume_size_in_gb=5,
)
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
f.update_profiler(rules=[profiler_custom_rule], system_monitor_interval_millis=1000)
sagemaker_session.update_training_job.assert_called_once()
_, args = sagemaker_session.update_training_job.call_args
assert args["profiler_config"] == {"ProfilingIntervalInMilliseconds": 1000}
assert args["profiler_rule_configs"] == [
{
"InstanceType": "c4.4xlarge",
"RuleConfigurationName": "CustomProfilerRule",
"RuleEvaluatorImage": "RuleImageUri",
"VolumeSizeInGB": 5,
}
]
def test_training_job_with_rule_job_summary(sagemaker_session, training_job_description):
f = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
f.fit("s3://mydata")
training_job_description["DebugRuleEvaluationStatuses"] = [
{
"RuleConfigurationName": "debugger_rule",
"RuleEvaluationJobArn": "debugger_rule_job_arn",
"RuleEvaluationStatus": "InProgress",
}
]
training_job_description["ProfilerRuleEvaluationStatuses"] = [
{
"RuleConfigurationName": "profiler_rule_1",
"RuleEvaluationJobArn": "profiler_rule_job_arn_1",
"RuleEvaluationStatus": "InProgress",
},
{
"RuleConfigurationName": "profiler_rule_2",
"RuleEvaluationJobArn": "profiler_rule_job_arn_2",
"RuleEvaluationStatus": "ERROR",
},
]
job_summary = f.latest_training_job.rule_job_summary()
assert job_summary == [
{
"RuleConfigurationName": "debugger_rule",
"RuleEvaluationJobArn": "debugger_rule_job_arn",
"RuleEvaluationStatus": "InProgress",
},
{
"RuleConfigurationName": "profiler_rule_1",
"RuleEvaluationJobArn": "profiler_rule_job_arn_1",
"RuleEvaluationStatus": "InProgress",
},
{
"RuleConfigurationName": "profiler_rule_2",
"RuleEvaluationJobArn": "profiler_rule_job_arn_2",
"RuleEvaluationStatus": "ERROR",
},
]
def test_framework_with_spot_and_checkpoints(sagemaker_session):
f = DummyFramework(
"my_script.py",
role="DummyRole",
instance_count=3,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
volume_size=123,
volume_kms_key="volumekms",
max_run=456,
input_mode="inputmode",
output_path="outputpath",
output_kms_key="outputkms",
base_job_name="basejobname",
tags=[{"foo": "bar"}],
subnets=["123", "456"],
security_group_ids=["789", "012"],
metric_definitions=[{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}],
encrypt_inter_container_traffic=True,
use_spot_instances=True,
max_wait=500,
checkpoint_s3_uri="s3://mybucket/checkpoints/",
checkpoint_local_path="/tmp/checkpoints",
)
_TrainingJob.start_new(f, "s3://mydata", None)
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args == {
"input_mode": "inputmode",
"tags": [{"foo": "bar"}],
"hyperparameters": {},
"image_uri": "fakeimage",
"input_config": [
{
"ChannelName": "training",
"DataSource": {
"S3DataSource": {
"S3DataType": "S3Prefix",
"S3DataDistributionType": "FullyReplicated",
"S3Uri": "s3://mydata",
}
},
}
],
"output_config": {"KmsKeyId": "outputkms", "S3OutputPath": "outputpath"},
"vpc_config": {"Subnets": ["123", "456"], "SecurityGroupIds": ["789", "012"]},
"stop_condition": {"MaxRuntimeInSeconds": 456, "MaxWaitTimeInSeconds": 500},
"role": sagemaker_session.expand_role(),
"job_name": None,
"resource_config": {
"VolumeSizeInGB": 123,
"InstanceCount": 3,
"VolumeKmsKeyId": "volumekms",
"InstanceType": "ml.m4.xlarge",
},
"metric_definitions": [{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}],
"encrypt_inter_container_traffic": True,
"use_spot_instances": True,
"checkpoint_s3_uri": "s3://mybucket/checkpoints/",
"checkpoint_local_path": "/tmp/checkpoints",
"environment": None,
"experiment_config": None,
"retry_strategy": None,
}
def test_framework_init_s3_entry_point_invalid(sagemaker_session):
with pytest.raises(ValueError) as error:
DummyFramework(
"s3://remote-script-because-im-mistaken",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
assert "Must be a path to a local file" in str(error)
def test_sagemaker_s3_uri_invalid(sagemaker_session):
with pytest.raises(ValueError) as error:
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
t.fit("thisdoesntstartwiths3")
assert "must be a valid S3 or FILE URI" in str(error)
def test_sagemaker_model_s3_uri_invalid(sagemaker_session):
with pytest.raises(ValueError) as error:
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
model_uri="thisdoesntstartwiths3either.tar.gz",
)
t.fit("s3://mydata")
assert "must be a valid S3 or FILE URI" in str(error)
def test_sagemaker_model_file_uri_invalid(sagemaker_session):
with pytest.raises(ValueError) as error:
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
model_uri="file://notins3.tar.gz",
)
t.fit("s3://mydata")
assert "File URIs are supported in local mode only" in str(error)
def test_sagemaker_model_default_channel_name(sagemaker_session):
f = DummyFramework(
entry_point="my_script.py",
role="DummyRole",
instance_count=3,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
model_uri="s3://model-bucket/prefix/model.tar.gz",
)
_TrainingJob.start_new(f, {}, None)
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["input_config"] == [
{
"ChannelName": "model",
"InputMode": "File",
"ContentType": "application/x-sagemaker-model",
"DataSource": {
"S3DataSource": {
"S3DataType": "S3Prefix",
"S3DataDistributionType": "FullyReplicated",
"S3Uri": "s3://model-bucket/prefix/model.tar.gz",
}
},
}
]
def test_sagemaker_model_custom_channel_name(sagemaker_session):
f = DummyFramework(
entry_point="my_script.py",
role="DummyRole",
instance_count=3,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
model_uri="s3://model-bucket/prefix/model.tar.gz",
model_channel_name="testModelChannel",
)
_TrainingJob.start_new(f, {}, None)
sagemaker_session.train.assert_called_once()
_, args = sagemaker_session.train.call_args
assert args["input_config"] == [
{
"ChannelName": "testModelChannel",
"InputMode": "File",
"ContentType": "application/x-sagemaker-model",
"DataSource": {
"S3DataSource": {
"S3DataType": "S3Prefix",
"S3DataDistributionType": "FullyReplicated",
"S3Uri": "s3://model-bucket/prefix/model.tar.gz",
}
},
}
]
@patch("time.strftime", return_value=TIMESTAMP)
def test_custom_code_bucket(time, sagemaker_session):
code_bucket = "codebucket"
prefix = "someprefix"
code_location = "s3://{}/{}".format(code_bucket, prefix)
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
code_location=code_location,
)
t.fit("s3://bucket/mydata")
expected_key = "{}/{}/source/sourcedir.tar.gz".format(prefix, JOB_NAME)
_, s3_args, _ = sagemaker_session.boto_session.resource("s3").Object.mock_calls[0]
assert s3_args == (code_bucket, expected_key)
expected_submit_dir = "s3://{}/{}".format(code_bucket, expected_key)
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
assert train_kwargs["hyperparameters"]["sagemaker_submit_directory"] == json.dumps(
expected_submit_dir
)
@patch("time.strftime", return_value=TIMESTAMP)
def test_custom_code_bucket_without_prefix(time, sagemaker_session):
code_bucket = "codebucket"
code_location = "s3://{}".format(code_bucket)
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
code_location=code_location,
)
t.fit("s3://bucket/mydata")
expected_key = "{}/source/sourcedir.tar.gz".format(JOB_NAME)
_, s3_args, _ = sagemaker_session.boto_session.resource("s3").Object.mock_calls[0]
assert s3_args == (code_bucket, expected_key)
expected_submit_dir = "s3://{}/{}".format(code_bucket, expected_key)
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
assert train_kwargs["hyperparameters"]["sagemaker_submit_directory"] == json.dumps(
expected_submit_dir
)
def test_invalid_custom_code_bucket(sagemaker_session):
code_location = "thisllworkright?"
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
code_location=code_location,
)
with pytest.raises(ValueError) as error:
t.fit("s3://bucket/mydata")
assert "Expecting 's3' scheme" in str(error)
def test_get_instance_type_gpu(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_groups=[
InstanceGroup("group1", "ml.c4.xlarge", 1),
InstanceGroup("group2", "ml.p3.16xlarge", 2),
],
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
assert "ml.p3.16xlarge" == estimator._get_instance_type()
def test_get_instance_type_cpu(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_groups=[
InstanceGroup("group1", "ml.c4.xlarge", 1),
InstanceGroup("group2", "ml.c5.xlarge", 2),
],
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
assert "ml.c4.xlarge" == estimator._get_instance_type()
def test_get_instance_type_no_instance_groups(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_type="ml.c4.xlarge",
instance_count=1,
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
assert "ml.c4.xlarge" == estimator._get_instance_type()
def test_get_instance_type_no_instance_groups_or_instance_type(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_type=None,
instance_count=None,
instance_groups=None,
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
with pytest.raises(ValueError) as error:
estimator._get_instance_type()
assert (
"instance_groups must be set if instance_type is not set and instance_groups must be a list."
in str(error)
)
def test_augmented_manifest(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit(
inputs=TrainingInput(
"s3://mybucket/train_manifest",
s3_data_type="AugmentedManifestFile",
attribute_names=["foo", "bar"],
)
)
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
s3_data_source = train_kwargs["input_config"][0]["DataSource"]["S3DataSource"]
assert s3_data_source["S3Uri"] == "s3://mybucket/train_manifest"
assert s3_data_source["S3DataType"] == "AugmentedManifestFile"
assert s3_data_source["AttributeNames"] == ["foo", "bar"]
def test_s3_input_mode(sagemaker_session):
expected_input_mode = "Pipe"
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit(inputs=TrainingInput("s3://mybucket/train_manifest", input_mode=expected_input_mode))
actual_input_mode = sagemaker_session.method_calls[1][2]["input_mode"]
assert actual_input_mode == expected_input_mode
def test_shuffle_config(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit(inputs=TrainingInput("s3://mybucket/train_manifest", shuffle_config=ShuffleConfig(100)))
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
channel = train_kwargs["input_config"][0]
assert channel["ShuffleConfig"]["Seed"] == 100
BASE_HP = {
"sagemaker_program": json.dumps(SCRIPT_NAME),
"sagemaker_submit_directory": json.dumps(
"s3://mybucket/{}/source/sourcedir.tar.gz".format(JOB_NAME)
),
"sagemaker_job_name": json.dumps(JOB_NAME),
}
def test_local_code_location():
config = {"local": {"local_code": True, "region": "us-west-2"}}
sms = Mock(
name="sagemaker_session",
boto_session=None,
boto_region_name=REGION,
config=config,
local_mode=True,
spec=sagemaker.local.LocalSession,
)
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sms,
instance_count=1,
instance_type="local",
base_job_name=IMAGE_URI,
hyperparameters={123: [456], "learning_rate": 0.1},
)
t.fit("file:///data/file")
assert t.source_dir == DATA_DIR
assert t.entry_point == "dummy_script.py"
@patch("time.strftime", return_value=TIMESTAMP)
def test_start_new_convert_hyperparameters_to_str(strftime, sagemaker_session):
uri = "bucket/mydata"
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
base_job_name=IMAGE_URI,
hyperparameters={123: [456], "learning_rate": 0.1},
)
t.fit("s3://{}".format(uri))
expected_hyperparameters = BASE_HP.copy()
expected_hyperparameters["sagemaker_container_log_level"] = str(logging.INFO)
expected_hyperparameters["learning_rate"] = json.dumps(0.1)
expected_hyperparameters["123"] = json.dumps([456])
expected_hyperparameters["sagemaker_region"] = '"us-west-2"'
actual_hyperparameter = sagemaker_session.method_calls[1][2]["hyperparameters"]
assert actual_hyperparameter == expected_hyperparameters
@patch("time.strftime", return_value=TIMESTAMP)
def test_start_new_wait_called(strftime, sagemaker_session):
uri = "bucket/mydata"
t = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
t.fit("s3://{}".format(uri))
expected_hyperparameters = BASE_HP.copy()
expected_hyperparameters["sagemaker_container_log_level"] = str(logging.INFO)
expected_hyperparameters["sagemaker_region"] = '"us-west-2"'
actual_hyperparameter = sagemaker_session.method_calls[1][2]["hyperparameters"]
assert actual_hyperparameter == expected_hyperparameters
assert sagemaker_session.wait_for_job.assert_called_once
def test_attach_framework(sagemaker_session, training_job_description):
training_job_description["VpcConfig"] = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
training_job_description["EnableNetworkIsolation"] = True
framework_estimator = DummyFramework.attach(
training_job_name="neo", sagemaker_session=sagemaker_session
)
assert framework_estimator._current_job_name == "neo"
assert framework_estimator.latest_training_job.job_name == "neo"
assert framework_estimator.role == "arn:aws:iam::366:role/SageMakerRole"
assert framework_estimator.instance_count == 1
assert framework_estimator.max_run == 24 * 60 * 60
assert framework_estimator.input_mode == "File"
assert framework_estimator.base_job_name == "neo"
assert framework_estimator.output_path == "s3://place/output/neo"
assert framework_estimator.output_kms_key == ""
assert framework_estimator.hyperparameters()["training_steps"] == "100"
assert framework_estimator.source_dir == "s3://some/sourcedir.tar.gz"
assert framework_estimator.entry_point == "iris-dnn-classifier.py"
assert framework_estimator.subnets == ["foo"]
assert framework_estimator.security_group_ids == ["bar"]
assert framework_estimator.encrypt_inter_container_traffic is False
assert framework_estimator.tags == LIST_TAGS_RESULT["Tags"]
assert framework_estimator.enable_network_isolation() is True
def test_attach_no_logs(sagemaker_session, training_job_description):
Estimator.attach(training_job_name="job", sagemaker_session=sagemaker_session)
sagemaker_session.logs_for_job.assert_not_called()
def test_logs(sagemaker_session, training_job_description):
estimator = Estimator.attach(training_job_name="job", sagemaker_session=sagemaker_session)
estimator.logs()
sagemaker_session.logs_for_job.assert_called_with(estimator.latest_training_job.name, wait=True)
def test_attach_without_hyperparameters(sagemaker_session, training_job_description):
del training_job_description["HyperParameters"]
estimator = Estimator.attach(training_job_name="job", sagemaker_session=sagemaker_session)
assert estimator.hyperparameters() == {}
def test_attach_framework_with_tuning(sagemaker_session, training_job_description):
training_job_description["HyperParameters"]["_tuning_objective_metric"] = "Validation-accuracy"
framework_estimator = DummyFramework.attach(
training_job_name="neo", sagemaker_session=sagemaker_session
)
assert framework_estimator.latest_training_job.job_name == "neo"
assert framework_estimator.role == "arn:aws:iam::366:role/SageMakerRole"
assert framework_estimator.instance_count == 1
assert framework_estimator.max_run == 24 * 60 * 60
assert framework_estimator.input_mode == "File"
assert framework_estimator.base_job_name == "neo"
assert framework_estimator.output_path == "s3://place/output/neo"
assert framework_estimator.output_kms_key == ""
hyper_params = framework_estimator.hyperparameters()
assert hyper_params["training_steps"] == "100"
assert hyper_params["_tuning_objective_metric"] == '"Validation-accuracy"'
assert framework_estimator.source_dir == "s3://some/sourcedir.tar.gz"
assert framework_estimator.entry_point == "iris-dnn-classifier.py"
assert framework_estimator.encrypt_inter_container_traffic is False
def test_attach_framework_with_model_channel(sagemaker_session, training_job_description):
s3_uri = "s3://some/s3/path/model.tar.gz"
training_job_description["InputDataConfig"] = [
{
"ChannelName": "model",
"InputMode": "File",
"DataSource": {"S3DataSource": {"S3Uri": s3_uri}},
}
]
framework_estimator = DummyFramework.attach(
training_job_name="neo", sagemaker_session=sagemaker_session
)
assert framework_estimator.model_uri is s3_uri
assert framework_estimator.encrypt_inter_container_traffic is False
def test_attach_framework_with_inter_container_traffic_encryption_flag(
sagemaker_session, training_job_description
):
training_job_description["EnableInterContainerTrafficEncryption"] = True
framework_estimator = DummyFramework.attach(
training_job_name="neo", sagemaker_session=sagemaker_session
)
assert framework_estimator.encrypt_inter_container_traffic is True
def test_attach_framework_base_from_generated_name(sagemaker_session, training_job_description):
base_job_name = "neo"
framework_estimator = DummyFramework.attach(
training_job_name=utils.name_from_base("neo"), sagemaker_session=sagemaker_session
)
assert framework_estimator.base_job_name == base_job_name
@patch("time.strftime", return_value=TIMESTAMP)
def test_fit_verify_job_name(strftime, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
tags=TAGS,
encrypt_inter_container_traffic=True,
)
fw.fit(inputs=TrainingInput("s3://mybucket/train"))
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
assert train_kwargs["image_uri"] == IMAGE_URI
assert train_kwargs["input_mode"] == "File"
assert train_kwargs["tags"] == TAGS
assert train_kwargs["job_name"] == JOB_NAME
assert train_kwargs["encrypt_inter_container_traffic"] is True
assert fw.latest_training_job.name == JOB_NAME
def test_prepare_for_training_unique_job_name_generation(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw._prepare_for_training()
first_job_name = fw._current_job_name
sleep(0.1)
fw._prepare_for_training()
second_job_name = fw._current_job_name
assert first_job_name != second_job_name
def test_prepare_for_training_force_name(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
base_job_name="some",
)
fw._prepare_for_training(job_name="use_it")
assert "use_it" == fw._current_job_name
@patch("time.strftime", return_value=TIMESTAMP)
def test_prepare_for_training_force_name_generation(strftime, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
base_job_name="some",
)
fw.base_job_name = None
fw._prepare_for_training()
assert JOB_NAME == fw._current_job_name
@patch("sagemaker.git_utils.git_clone_repo")
def test_git_support_with_branch_and_commit_succeed(git_clone_repo, sagemaker_session):
git_clone_repo.side_effect = lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": [],
}
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
@patch("sagemaker.git_utils.git_clone_repo")
def test_git_support_with_branch_succeed(git_clone_repo, sagemaker_session):
git_clone_repo.side_effect = lambda gitconfig, entrypoint, source_dir, dependencies=None: {
"entry_point": "/tmp/repo_dir/source_dir/entry_point",
"source_dir": None,
"dependencies": None,
}
git_config = {"repo": GIT_REPO, "branch": BRANCH}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
@patch("sagemaker.git_utils.git_clone_repo")
def test_git_support_with_dependencies_succeed(git_clone_repo, sagemaker_session):
git_clone_repo.side_effect = lambda gitconfig, entrypoint, source_dir, dependencies: {
"entry_point": "/tmp/repo_dir/source_dir/entry_point",
"source_dir": None,
"dependencies": ["/tmp/repo_dir/foo", "/tmp/repo_dir/foo/bar"],
}
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
entry_point = "source_dir/entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
dependencies=["foo", "foo/bar"],
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, ["foo", "foo/bar"])
@patch("sagemaker.git_utils.git_clone_repo")
def test_git_support_without_branch_and_commit_succeed(git_clone_repo, sagemaker_session):
git_clone_repo.side_effect = lambda gitconfig, entrypoint, source_dir, dependencies=None: {
"entry_point": "/tmp/repo_dir/source_dir/entry_point",
"source_dir": None,
"dependencies": None,
}
git_config = {"repo": GIT_REPO}
entry_point = "source_dir/entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
def test_git_support_repo_not_provided(sagemaker_session):
git_config = {"branch": BRANCH, "commit": COMMIT}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
source_dir="source_dir",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(ValueError) as error:
fw.fit()
assert "Please provide a repo for git_config." in str(error)
def test_git_support_bad_repo_url_format(sagemaker_session):
git_config = {"repo": "hhttps://github.com/user/repo.git", "branch": BRANCH}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
source_dir="source_dir",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(ValueError) as error:
fw.fit()
assert "Invalid Git url provided." in str(error)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=subprocess.CalledProcessError(
returncode=1, cmd="git clone https://github.com/aws/no-such-repo.git /tmp/repo_dir"
),
)
def test_git_support_git_clone_fail(git_clone_repo, sagemaker_session):
git_config = {"repo": "https://github.com/aws/no-such-repo.git", "branch": BRANCH}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(subprocess.CalledProcessError) as error:
fw.fit()
assert "returned non-zero exit status" in str(error.value)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=subprocess.CalledProcessError(
returncode=1, cmd="git checkout branch-that-does-not-exist"
),
)
def test_git_support_branch_not_exist(git_clone_repo, sagemaker_session):
git_config = {"repo": GIT_REPO, "branch": "branch-that-does-not-exist", "commit": COMMIT}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(subprocess.CalledProcessError) as error:
fw.fit()
assert "returned non-zero exit status" in str(error.value)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=subprocess.CalledProcessError(
returncode=1, cmd="git checkout commit-sha-that-does-not-exist"
),
)
def test_git_support_commit_not_exist(git_clone_repo, sagemaker_session):
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": "commit-sha-that-does-not-exist"}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(subprocess.CalledProcessError) as error:
fw.fit()
assert "returned non-zero exit status" in str(error.value)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=ValueError("Entry point does not exist in the repo."),
)
def test_git_support_entry_point_not_exist(git_clone_repo, sagemaker_session):
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
fw = DummyFramework(
entry_point="entry_point_that_does_not_exist",
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(ValueError) as error:
fw.fit()
assert "Entry point does not exist in the repo." in str(error)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=ValueError("Source directory does not exist in the repo."),
)
def test_git_support_source_dir_not_exist(git_clone_repo, sagemaker_session):
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
source_dir="source_dir_that_does_not_exist",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(ValueError) as error:
fw.fit()
assert "Source directory does not exist in the repo." in str(error)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=ValueError("Dependency no-such-dir does not exist in the repo."),
)
def test_git_support_dependencies_not_exist(git_clone_repo, sagemaker_session):
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
fw = DummyFramework(
entry_point="entry_point",
git_config=git_config,
source_dir="source_dir",
dependencies=["foo", "no-such-dir"],
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(ValueError) as error:
fw.fit()
assert "Dependency", "does not exist in the repo." in str(error)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
},
)
def test_git_support_with_username_password_no_2fa(git_clone_repo, sagemaker_session):
git_config = {
"repo": PRIVATE_GIT_REPO,
"branch": PRIVATE_BRANCH,
"commit": PRIVATE_COMMIT,
"username": "username",
"password": "passw0rd!",
}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
assert fw.entry_point == "/tmp/repo_dir/entry_point"
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
},
)
def test_git_support_with_token_2fa(git_clone_repo, sagemaker_session):
git_config = {
"repo": PRIVATE_GIT_REPO,
"branch": PRIVATE_BRANCH,
"commit": PRIVATE_COMMIT,
"token": "my-token",
"2FA_enabled": True,
}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
assert fw.entry_point == "/tmp/repo_dir/entry_point"
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
},
)
def test_git_support_ssh_no_passphrase_needed(git_clone_repo, sagemaker_session):
git_config = {"repo": PRIVATE_GIT_REPO_SSH, "branch": PRIVATE_BRANCH, "commit": PRIVATE_COMMIT}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
assert fw.entry_point == "/tmp/repo_dir/entry_point"
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=subprocess.CalledProcessError(
returncode=1, cmd="git clone {} {}".format(PRIVATE_GIT_REPO_SSH, REPO_DIR)
),
)
def test_git_support_ssh_passphrase_required(git_clone_repo, sagemaker_session):
git_config = {"repo": PRIVATE_GIT_REPO_SSH, "branch": PRIVATE_BRANCH, "commit": PRIVATE_COMMIT}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
with pytest.raises(subprocess.CalledProcessError) as error:
fw.fit()
assert "returned non-zero exit status" in str(error.value)
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
},
)
def test_git_support_codecommit_with_username_and_password_succeed(
git_clone_repo, sagemaker_session
):
git_config = {
"repo": CODECOMMIT_REPO,
"branch": CODECOMMIT_BRANCH,
"username": "username",
"password": "passw0rd!",
}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
assert fw.entry_point == "/tmp/repo_dir/entry_point"
@patch(
"sagemaker.git_utils.git_clone_repo",
side_effect=lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
},
)
def test_git_support_codecommit_with_ssh_no_passphrase_needed(git_clone_repo, sagemaker_session):
git_config = {"repo": CODECOMMIT_REPO_SSH, "branch": CODECOMMIT_BRANCH}
entry_point = "entry_point"
fw = DummyFramework(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
assert fw.entry_point == "/tmp/repo_dir/entry_point"
@patch("time.strftime", return_value=TIMESTAMP)
def test_init_with_source_dir_s3(strftime, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_NAME,
source_dir="s3://location",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
fw._prepare_for_training()
expected_hyperparameters = {
"sagemaker_program": SCRIPT_NAME,
"sagemaker_job_name": JOB_NAME,
"sagemaker_container_log_level": logging.INFO,
"sagemaker_submit_directory": "s3://location",
"sagemaker_region": "us-west-2",
}
assert fw._hyperparameters == expected_hyperparameters
@patch("sagemaker.estimator.name_from_base", return_value=MODEL_IMAGE)
def test_framework_transformer_creation(name_from_base, sagemaker_session):
vpc_config = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
subnets=vpc_config["Subnets"],
security_group_ids=vpc_config["SecurityGroupIds"],
)
fw.latest_training_job = _TrainingJob(sagemaker_session, JOB_NAME)
transformer = fw.transformer(INSTANCE_COUNT, INSTANCE_TYPE)
name_from_base.assert_called_with(IMAGE_URI)
sagemaker_session.create_model.assert_called_with(
name=MODEL_IMAGE,
role=ROLE,
container_defs=MODEL_CONTAINER_DEF,
tags=None,
vpc_config=vpc_config,
enable_network_isolation=False,
)
assert isinstance(transformer, Transformer)
assert transformer.sagemaker_session == sagemaker_session
assert transformer.instance_count == INSTANCE_COUNT
assert transformer.instance_type == INSTANCE_TYPE
assert transformer.model_name == MODEL_IMAGE
assert transformer.tags is None
assert transformer.env == {}
@patch("sagemaker.model.utils.name_from_image", return_value=MODEL_IMAGE)
def test_framework_transformer_creation_with_optional_params(name_from_image, sagemaker_session):
base_name = "foo"
vpc_config = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
base_job_name=base_name,
subnets=vpc_config["Subnets"],
security_group_ids=vpc_config["SecurityGroupIds"],
enable_network_isolation=False,
)
fw.latest_training_job = _TrainingJob(sagemaker_session, JOB_NAME)
strategy = "MultiRecord"
assemble_with = "Line"
kms_key = "key"
accept = "text/csv"
max_concurrent_transforms = 1
max_payload = 6
env = {"FOO": "BAR"}
new_role = "dummy-model-role"
new_vpc_config = {"Subnets": ["x"], "SecurityGroupIds": ["y"]}
model_name = "model-name"
transformer = fw.transformer(
INSTANCE_COUNT,
INSTANCE_TYPE,
strategy=strategy,
assemble_with=assemble_with,
output_path=OUTPUT_PATH,
output_kms_key=kms_key,
accept=accept,
tags=TAGS,
max_concurrent_transforms=max_concurrent_transforms,
max_payload=max_payload,
volume_kms_key=kms_key,
env=env,
role=new_role,
model_server_workers=1,
vpc_config_override=new_vpc_config,
enable_network_isolation=True,
model_name=model_name,
)
sagemaker_session.create_model.assert_called_with(
name=model_name,
role=new_role,
container_defs=MODEL_CONTAINER_DEF,
vpc_config=new_vpc_config,
tags=TAGS,
enable_network_isolation=True,
)
assert transformer.strategy == strategy
assert transformer.assemble_with == assemble_with
assert transformer.output_path == OUTPUT_PATH
assert transformer.output_kms_key == kms_key
assert transformer.accept == accept
assert transformer.max_concurrent_transforms == max_concurrent_transforms
assert transformer.max_payload == max_payload
assert transformer.env == env
assert transformer.base_transform_job_name == base_name
assert transformer.tags == TAGS
assert transformer.volume_kms_key == kms_key
assert transformer.model_name == model_name
def test_ensure_latest_training_job(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
)
fw.latest_training_job = Mock(name="training_job")
fw._ensure_latest_training_job()
def test_ensure_latest_training_job_failure(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
)
with pytest.raises(ValueError) as e:
fw._ensure_latest_training_job()
assert "Estimator is not associated with a training job" in str(e)
@patch("sagemaker.estimator.Estimator.create_model")
@patch("sagemaker.estimator.name_from_base")
def test_estimator_transformer_creation(name_from_base, create_model, sagemaker_session):
estimator = Estimator(
image_uri=IMAGE_URI,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
)
estimator.latest_training_job = _TrainingJob(sagemaker_session, JOB_NAME)
model_name = "model_name"
name_from_base.return_value = model_name
transformer = estimator.transformer(INSTANCE_COUNT, INSTANCE_TYPE)
create_model.assert_called_with(
vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT,
model_kms_key=estimator.output_kms_key,
enable_network_isolation=False,
)
assert isinstance(transformer, Transformer)
assert transformer.sagemaker_session == sagemaker_session
assert transformer.instance_count == INSTANCE_COUNT
assert transformer.instance_type == INSTANCE_TYPE
assert transformer.model_name == model_name
assert transformer.tags is None
@patch("sagemaker.estimator.Estimator.create_model")
def test_estimator_transformer_creation_with_optional_params(create_model, sagemaker_session):
base_name = "foo"
kms_key = "key"
estimator = Estimator(
image_uri=IMAGE_URI,
role=ROLE,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
sagemaker_session=sagemaker_session,
base_job_name=base_name,
output_kms_key=kms_key,
)
estimator.latest_training_job = _TrainingJob(sagemaker_session, JOB_NAME)
strategy = "MultiRecord"
assemble_with = "Line"
accept = "text/csv"
max_concurrent_transforms = 1
max_payload = 6
env = {"FOO": "BAR"}
new_vpc_config = {"Subnets": ["x"], "SecurityGroupIds": ["y"]}
model_name = "model-name"
transformer = estimator.transformer(
INSTANCE_COUNT,
INSTANCE_TYPE,
strategy=strategy,
assemble_with=assemble_with,
output_path=OUTPUT_PATH,
output_kms_key=kms_key,
accept=accept,
tags=TAGS,
max_concurrent_transforms=max_concurrent_transforms,
max_payload=max_payload,
env=env,
role=ROLE,
vpc_config_override=new_vpc_config,
enable_network_isolation=True,
model_name=model_name,
)
create_model.assert_called_with(
vpc_config_override=new_vpc_config, model_kms_key=kms_key, enable_network_isolation=True
)
assert transformer.strategy == strategy
assert transformer.assemble_with == assemble_with
assert transformer.output_path == OUTPUT_PATH
assert transformer.output_kms_key == kms_key
assert transformer.accept == accept
assert transformer.max_concurrent_transforms == max_concurrent_transforms
assert transformer.max_payload == max_payload
assert transformer.env == env
assert transformer.base_transform_job_name == base_name
assert transformer.tags == TAGS
assert transformer.model_name == model_name
# _TrainingJob 'utils'
def test_start_new(sagemaker_session):
training_job = _TrainingJob(sagemaker_session, JOB_NAME)
hyperparameters = {"mock": "hyperparameters"}
inputs = "s3://mybucket/train"
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
hyperparameters=hyperparameters,
)
exp_config = {"ExperimentName": "exp", "TrialName": "t", "TrialComponentDisplayName": "tc"}
started_training_job = training_job.start_new(estimator, inputs, experiment_config=exp_config)
called_args = sagemaker_session.train.call_args
assert started_training_job.sagemaker_session == sagemaker_session
assert called_args[1]["hyperparameters"] == hyperparameters
assert called_args[1]["experiment_config"] == exp_config
sagemaker_session.train.assert_called_once()
def test_start_new_not_local_mode_error(sagemaker_session):
training_job = _TrainingJob(sagemaker_session, JOB_NAME)
inputs = "file://mybucket/train"
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
with pytest.raises(ValueError) as error:
training_job.start_new(estimator, inputs, None)
assert "File URIs are supported in local mode only. Please use a S3 URI instead." == str(
error
)
def test_container_log_level(sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
container_log_level=logging.DEBUG,
)
fw.fit(inputs=TrainingInput("s3://mybucket/train"))
_, _, train_kwargs = sagemaker_session.train.mock_calls[0]
assert train_kwargs["hyperparameters"]["sagemaker_container_log_level"] == "10"
@patch("sagemaker.utils")
def test_same_code_location_keeps_kms_key(utils, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
output_kms_key="kms-key",
)
fw.fit(wait=False)
extra_args = {"ServerSideEncryption": "aws:kms", "SSEKMSKeyId": "kms-key"}
obj = sagemaker_session.boto_session.resource("s3").Object
obj.assert_called_with("mybucket", "%s/source/sourcedir.tar.gz" % fw._current_job_name)
obj().upload_file.assert_called_with(utils.create_tar_file(), ExtraArgs=extra_args)
@patch("sagemaker.utils")
def test_different_code_location_kms_key(utils, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
code_location="s3://another-location",
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
output_kms_key="kms-key",
)
fw.fit(wait=False)
obj = sagemaker_session.boto_session.resource("s3").Object
obj.assert_called_with("another-location", "%s/source/sourcedir.tar.gz" % fw._current_job_name)
extra_args = {"ServerSideEncryption": "aws:kms"}
obj().upload_file.assert_called_with(utils.create_tar_file(), ExtraArgs=extra_args)
@patch("sagemaker.utils")
def test_default_code_location_uses_output_path(utils, sagemaker_session):
fw = DummyFramework(
entry_point=SCRIPT_PATH,
role="DummyRole",
sagemaker_session=sagemaker_session,
output_path="s3://output_path",
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
output_kms_key="kms-key",
)
fw.fit(wait=False)
obj = sagemaker_session.boto_session.resource("s3").Object
obj.assert_called_with("output_path", "%s/source/sourcedir.tar.gz" % fw._current_job_name)
extra_args = {"ServerSideEncryption": "aws:kms", "SSEKMSKeyId": "kms-key"}
obj().upload_file.assert_called_with(utils.create_tar_file(), ExtraArgs=extra_args)
def test_wait_without_logs(sagemaker_session):
training_job = _TrainingJob(sagemaker_session, JOB_NAME)
training_job.wait(False)
sagemaker_session.wait_for_job.assert_called_once()
assert not sagemaker_session.logs_for_job.called
def test_wait_with_logs(sagemaker_session):
training_job = _TrainingJob(sagemaker_session, JOB_NAME)
training_job.wait()
sagemaker_session.logs_for_job.assert_called_once()
assert not sagemaker_session.wait_for_job.called
def test_unsupported_type_in_dict():
with pytest.raises(ValueError):
_TrainingJob._format_inputs_to_input_config({"a": 66})
#################################################################################
# Tests for the generic Estimator class
NO_INPUT_TRAIN_CALL = {
"hyperparameters": {},
"image_uri": IMAGE_URI,
"input_config": None,
"input_mode": "File",
"output_config": {"S3OutputPath": OUTPUT_PATH},
"profiler_config": {"S3OutputPath": OUTPUT_PATH},
"profiler_rule_configs": [
{
"RuleConfigurationName": "ProfilerReport-1510006209",
"RuleEvaluatorImage": "895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest",
"RuleParameters": {"rule_to_invoke": "ProfilerReport"},
}
],
"resource_config": {
"InstanceCount": INSTANCE_COUNT,
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": 30,
},
"stop_condition": {"MaxRuntimeInSeconds": 86400},
"retry_strategy": None,
"tags": None,
"vpc_config": None,
"metric_definitions": None,
"environment": None,
"experiment_config": None,
}
INPUT_CONFIG = [
{
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": "s3://bucket/training-prefix",
}
},
"ChannelName": "train",
}
]
BASE_TRAIN_CALL = dict(NO_INPUT_TRAIN_CALL)
BASE_TRAIN_CALL.update({"input_config": INPUT_CONFIG})
HYPERPARAMS = {"x": 1, "y": "hello"}
STRINGIFIED_HYPERPARAMS = dict([(x, str(y)) for x, y in HYPERPARAMS.items()])
HP_TRAIN_CALL = dict(BASE_TRAIN_CALL)
HP_TRAIN_CALL.update({"hyperparameters": STRINGIFIED_HYPERPARAMS})
EXP_TRAIN_CALL = dict(BASE_TRAIN_CALL)
EXP_TRAIN_CALL.update(
{
"experiment_config": {
"ExperimentName": "exp",
"TrialName": "trial",
"TrialComponentDisplayName": "tc",
}
}
)
@patch("sagemaker.estimator.name_from_base")
def test_fit_deploy_tags_in_estimator(name_from_base, sagemaker_session):
tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}]
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
tags=tags,
sagemaker_session=sagemaker_session,
)
estimator.fit()
model_name = "model_name"
name_from_base.return_value = model_name
estimator.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
variant = [
{
"InstanceType": "c4.4xlarge",
"VariantName": "AllTraffic",
"ModelName": model_name,
"InitialVariantWeight": 1,
"InitialInstanceCount": 1,
}
]
name_from_base.assert_called_with(IMAGE_URI)
sagemaker_session.endpoint_from_production_variants.assert_called_with(
name=model_name,
production_variants=variant,
tags=tags,
kms_key=None,
wait=True,
data_capture_config_dict=None,
async_inference_config_dict=None,
)
sagemaker_session.create_model.assert_called_with(
name=model_name,
role="DummyRole",
container_defs={
"ModelDataUrl": "s3://bucket/model.tar.gz",
"Environment": {},
"Image": "fakeimage",
},
enable_network_isolation=False,
vpc_config=None,
tags=tags,
)
@patch("sagemaker.estimator.name_from_base")
def test_fit_deploy_tags(name_from_base, sagemaker_session):
estimator = Estimator(
IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE, sagemaker_session=sagemaker_session
)
estimator.fit()
model_name = "model_name"
name_from_base.return_value = model_name
tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}]
estimator.deploy(INSTANCE_COUNT, INSTANCE_TYPE, tags=tags)
variant = [
{
"InstanceType": "c4.4xlarge",
"VariantName": "AllTraffic",
"ModelName": model_name,
"InitialVariantWeight": 1,
"InitialInstanceCount": 1,
}
]
name_from_base.assert_called_with(IMAGE_URI)
sagemaker_session.endpoint_from_production_variants.assert_called_with(
name=model_name,
production_variants=variant,
tags=tags,
kms_key=None,
wait=True,
data_capture_config_dict=None,
async_inference_config_dict=None,
)
sagemaker_session.create_model.assert_called_with(
name=ANY,
role="DummyRole",
container_defs={
"ModelDataUrl": "s3://bucket/model.tar.gz",
"Environment": {},
"Image": "fakeimage",
},
enable_network_isolation=False,
vpc_config=None,
tags=tags,
)
@patch("time.time", return_value=TIME)
def test_generic_to_fit_no_input(time, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit()
sagemaker_session.train.assert_called_once()
assert len(sagemaker_session.train.call_args[0]) == 0
args = sagemaker_session.train.call_args[1]
assert args["job_name"].startswith(IMAGE_URI)
args.pop("job_name")
args.pop("role")
args.pop("debugger_hook_config")
assert args == NO_INPUT_TRAIN_CALL
@patch("time.time", return_value=TIME)
def test_generic_to_fit_no_hps(time, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit({"train": "s3://bucket/training-prefix"})
sagemaker_session.train.assert_called_once()
assert len(sagemaker_session.train.call_args[0]) == 0
args = sagemaker_session.train.call_args[1]
assert args["job_name"].startswith(IMAGE_URI)
args.pop("job_name")
args.pop("role")
args.pop("debugger_hook_config")
assert args == BASE_TRAIN_CALL
@patch("time.time", return_value=TIME)
def test_generic_to_fit_with_hps(time, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.set_hyperparameters(**HYPERPARAMS)
e.fit({"train": "s3://bucket/training-prefix"})
sagemaker_session.train.assert_called_once()
assert len(sagemaker_session.train.call_args[0]) == 0
args = sagemaker_session.train.call_args[1]
assert args["job_name"].startswith(IMAGE_URI)
args.pop("job_name")
args.pop("role")
args.pop("debugger_hook_config")
assert args == HP_TRAIN_CALL
@patch("time.time", return_value=TIME)
def test_generic_to_fit_with_experiment_config(time, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit(
inputs={"train": "s3://bucket/training-prefix"},
experiment_config={
"ExperimentName": "exp",
"TrialName": "trial",
"TrialComponentDisplayName": "tc",
},
)
sagemaker_session.train.assert_called_once()
assert len(sagemaker_session.train.call_args[0]) == 0
args = sagemaker_session.train.call_args[1]
assert args["job_name"].startswith(IMAGE_URI)
args.pop("job_name")
args.pop("role")
args.pop("debugger_hook_config")
assert args == EXP_TRAIN_CALL
def test_generic_to_fit_with_encrypt_inter_container_traffic_flag(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
encrypt_inter_container_traffic=True,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert args["encrypt_inter_container_traffic"] is True
def test_generic_to_fit_with_network_isolation(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
enable_network_isolation=True,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert args["enable_network_isolation"]
def test_generic_to_fit_with_sagemaker_metrics_missing(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert "enable_sagemaker_metrics" not in args
def test_add_environment_variables_to_train_args(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
environment=ENV_INPUT,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert args["environment"] == ENV_INPUT
def test_add_retry_strategy_to_train_args(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
max_retry_attempts=2,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert args["retry_strategy"] == {"MaximumRetryAttempts": 2}
def test_generic_to_fit_with_sagemaker_metrics_enabled(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
enable_sagemaker_metrics=True,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert args["enable_sagemaker_metrics"]
def test_generic_to_fit_with_sagemaker_metrics_disabled(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
enable_sagemaker_metrics=False,
)
e.fit()
sagemaker_session.train.assert_called_once()
args = sagemaker_session.train.call_args[1]
assert not args["enable_sagemaker_metrics"]
@patch("time.time", return_value=TIME)
def test_generic_to_deploy(time, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.set_hyperparameters(**HYPERPARAMS)
e.fit({"train": "s3://bucket/training-prefix"})
predictor = e.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
sagemaker_session.train.assert_called_once()
assert len(sagemaker_session.train.call_args[0]) == 0
args = sagemaker_session.train.call_args[1]
assert args["job_name"].startswith(IMAGE_URI)
args.pop("job_name")
args.pop("role")
args.pop("debugger_hook_config")
assert args == HP_TRAIN_CALL
sagemaker_session.create_model.assert_called_once()
args, kwargs = sagemaker_session.create_model.call_args
assert kwargs["name"].startswith(IMAGE_URI)
assert kwargs["role"] == ROLE
assert kwargs["container_defs"]["Image"] == IMAGE_URI
assert kwargs["container_defs"]["ModelDataUrl"] == MODEL_DATA
assert kwargs["vpc_config"] is None
assert isinstance(predictor, Predictor)
assert predictor.endpoint_name.startswith(IMAGE_URI)
assert predictor.sagemaker_session == sagemaker_session
def test_generic_to_deploy_async(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit()
s3_output_path = "s3://some-s3-path"
predictor_async = e.deploy(
INSTANCE_COUNT,
INSTANCE_TYPE,
async_inference_config=AsyncInferenceConfig(output_path=s3_output_path),
)
sagemaker_session.create_model.assert_called_once()
_, kwargs = sagemaker_session.create_model.call_args
assert isinstance(predictor_async, AsyncPredictor)
assert predictor_async.endpoint_name.startswith(IMAGE_URI)
assert predictor_async.sagemaker_session == sagemaker_session
def test_generic_to_deploy_bad_arguments_combination(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit()
bad_args = (
{"instance_type": INSTANCE_TYPE},
{"initial_instance_count": INSTANCE_COUNT},
{"instance_type": None, "initial_instance_count": None},
)
for args in bad_args:
with pytest.raises(
ValueError,
match="Must specify instance type and instance count unless using serverless inference",
):
e.deploy(args)
with pytest.raises(
ValueError,
match="serverless_inference_config needs to be a ServerlessInferenceConfig object",
):
e.deploy(serverless_inference_config={})
def test_generic_to_deploy_network_isolation(sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
enable_network_isolation=True,
sagemaker_session=sagemaker_session,
)
e.fit()
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
sagemaker_session.create_model.assert_called_once()
_, kwargs = sagemaker_session.create_model.call_args
assert kwargs["enable_network_isolation"]
@patch("sagemaker.estimator.Estimator.create_model")
def test_generic_to_deploy_kms(create_model, sagemaker_session):
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
e.fit()
model = MagicMock()
create_model.return_value = model
endpoint_name = "foo"
kms_key = "key"
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE, endpoint_name=endpoint_name, kms_key=kms_key)
model.deploy.assert_called_with(
instance_type=INSTANCE_TYPE,
initial_instance_count=INSTANCE_COUNT,
serializer=None,
deserializer=None,
accelerator_type=None,
endpoint_name=endpoint_name,
tags=None,
wait=True,
kms_key=kms_key,
data_capture_config=None,
async_inference_config=None,
serverless_inference_config=None,
)
def test_generic_training_job_analytics(sagemaker_session):
sagemaker_session.sagemaker_client.describe_training_job = Mock(
name="describe_training_job",
return_value={
"TuningJobArn": "arn:aws:sagemaker:us-west-2:968277160000:hyper-parameter-tuning-job/mock-tuner",
"TrainingStartTime": 1530562991.299,
"AlgorithmSpecification": {
"TrainingImage": "some-image-url",
"TrainingInputMode": "File",
"MetricDefinitions": [
{"Name": "train:loss", "Regex": "train_loss=([0-9]+\\.[0-9]+)"},
{"Name": "validation:loss", "Regex": "valid_loss=([0-9]+\\.[0-9]+)"},
],
},
},
)
e = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
with pytest.raises(ValueError) as err: # noqa: F841
# No training job yet
a = e.training_job_analytics
assert a is not None # This line is never reached
e.set_hyperparameters(**HYPERPARAMS)
e.fit({"train": "s3://bucket/training-prefix"})
a = e.training_job_analytics
assert a is not None
def test_generic_create_model_vpc_config_override(sagemaker_session):
vpc_config_a = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
vpc_config_b = {"Subnets": ["foo", "bar"], "SecurityGroupIds": ["baz"]}
e = Estimator(
IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE, sagemaker_session=sagemaker_session
)
e.fit({"train": "s3://bucket/training-prefix"})
assert e.get_vpc_config() is None
assert e.create_model().vpc_config is None
assert e.create_model(vpc_config_override=vpc_config_a).vpc_config == vpc_config_a
assert e.create_model(vpc_config_override=None).vpc_config is None
e.subnets = vpc_config_a["Subnets"]
e.security_group_ids = vpc_config_a["SecurityGroupIds"]
assert e.get_vpc_config() == vpc_config_a
assert e.create_model().vpc_config == vpc_config_a
assert e.create_model(vpc_config_override=vpc_config_b).vpc_config == vpc_config_b
assert e.create_model(vpc_config_override=None).vpc_config is None
with pytest.raises(ValueError):
e.get_vpc_config(vpc_config_override={"invalid"})
with pytest.raises(ValueError):
e.create_model(vpc_config_override={"invalid"})
def test_generic_deploy_vpc_config_override(sagemaker_session):
vpc_config_a = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
vpc_config_b = {"Subnets": ["foo", "bar"], "SecurityGroupIds": ["baz"]}
e = Estimator(
IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE, sagemaker_session=sagemaker_session
)
e.fit({"train": "s3://bucket/training-prefix"})
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
assert sagemaker_session.create_model.call_args_list[0][1]["vpc_config"] is None
e.subnets = vpc_config_a["Subnets"]
e.security_group_ids = vpc_config_a["SecurityGroupIds"]
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
assert sagemaker_session.create_model.call_args_list[1][1]["vpc_config"] == vpc_config_a
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE, vpc_config_override=vpc_config_b)
assert sagemaker_session.create_model.call_args_list[2][1]["vpc_config"] == vpc_config_b
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE, vpc_config_override=None)
assert sagemaker_session.create_model.call_args_list[3][1]["vpc_config"] is None
def test_generic_deploy_accelerator_type(sagemaker_session):
e = Estimator(
IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE, sagemaker_session=sagemaker_session
)
e.fit({"train": "s3://bucket/training-prefix"})
e.deploy(INSTANCE_COUNT, INSTANCE_TYPE, accelerator_type=ACCELERATOR_TYPE)
args = e.sagemaker_session.endpoint_from_production_variants.call_args[1]
assert args["name"].startswith(IMAGE_URI)
assert args["production_variants"][0]["AcceleratorType"] == ACCELERATOR_TYPE
assert args["production_variants"][0]["InitialInstanceCount"] == INSTANCE_COUNT
assert args["production_variants"][0]["InstanceType"] == INSTANCE_TYPE
def test_deploy_with_model_name(sagemaker_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
estimator.set_hyperparameters(**HYPERPARAMS)
estimator.fit({"train": "s3://bucket/training-prefix"})
model_name = "model-name"
estimator.deploy(INSTANCE_COUNT, INSTANCE_TYPE, model_name=model_name)
sagemaker_session.create_model.assert_called_once()
args, kwargs = sagemaker_session.create_model.call_args
assert kwargs["name"] == model_name
def test_deploy_with_no_model_name(sagemaker_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
estimator.set_hyperparameters(**HYPERPARAMS)
estimator.fit({"train": "s3://bucket/training-prefix"})
estimator.deploy(INSTANCE_COUNT, INSTANCE_TYPE)
sagemaker_session.create_model.assert_called_once()
args, kwargs = sagemaker_session.create_model.call_args
assert kwargs["name"].startswith(IMAGE_URI)
def test_register_default_image(sagemaker_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
estimator.set_hyperparameters(**HYPERPARAMS)
estimator.fit({"train": "s3://bucket/training-prefix"})
model_package_name = "test-estimator-register-model"
content_types = ["application/json"]
response_types = ["application/json"]
inference_instances = ["ml.m4.xlarge"]
transform_instances = ["ml.m4.xlarget"]
sample_payload_url = "s3://test-bucket/model"
task = "IMAGE_CLASSIFICATION"
framework = "TENSORFLOW"
framework_version = "2.9"
nearest_model_name = "resnet50"
data_input_config = '{"input_1":[1,224,224,3]}'
estimator.register(
content_types=content_types,
response_types=response_types,
inference_instances=inference_instances,
transform_instances=transform_instances,
model_package_name=model_package_name,
sample_payload_url=sample_payload_url,
task=task,
framework=framework,
framework_version=framework_version,
nearest_model_name=nearest_model_name,
data_input_configuration=data_input_config,
)
sagemaker_session.create_model.assert_not_called()
expected_create_model_package_request = {
"containers": [
{
"Image": estimator.image_uri,
"ModelDataUrl": estimator.model_data,
}
],
"content_types": content_types,
"response_types": response_types,
"inference_instances": inference_instances,
"transform_instances": transform_instances,
"model_package_name": model_package_name,
"marketplace_cert": False,
"sample_payload_url": sample_payload_url,
"task": task,
}
sagemaker_session.create_model_package_from_containers.assert_called_with(
**expected_create_model_package_request
)
def test_register_default_image_without_instance_type_args(sagemaker_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
estimator.set_hyperparameters(**HYPERPARAMS)
estimator.fit({"train": "s3://bucket/training-prefix"})
model_package_name = "test-estimator-register-model"
content_types = ["application/json"]
response_types = ["application/json"]
sample_payload_url = "s3://test-bucket/model"
task = "IMAGE_CLASSIFICATION"
framework = "TENSORFLOW"
framework_version = "2.9"
nearest_model_name = "resnet50"
estimator.register(
content_types=content_types,
response_types=response_types,
model_package_name=model_package_name,
sample_payload_url=sample_payload_url,
task=task,
framework=framework,
framework_version=framework_version,
nearest_model_name=nearest_model_name,
)
sagemaker_session.create_model.assert_not_called()
expected_create_model_package_request = {
"containers": [
{
"Image": estimator.image_uri,
"ModelDataUrl": estimator.model_data,
}
],
"content_types": content_types,
"response_types": response_types,
"inference_instances": None,
"transform_instances": None,
"model_package_name": model_package_name,
"marketplace_cert": False,
"sample_payload_url": sample_payload_url,
"task": task,
}
sagemaker_session.create_model_package_from_containers.assert_called_with(
**expected_create_model_package_request
)
def test_register_inference_image(sagemaker_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=sagemaker_session,
)
estimator.set_hyperparameters(**HYPERPARAMS)
estimator.fit({"train": "s3://bucket/training-prefix"})
model_package_name = "test-estimator-register-model"
content_types = ["application/json"]
response_types = ["application/json"]
inference_instances = ["ml.m4.xlarge"]
transform_instances = ["ml.m4.xlarget"]
inference_image = "fake-inference-image"
sample_payload_url = "s3://test-bucket/model"
task = "IMAGE_CLASSIFICATION"
framework = "TENSORFLOW"
framework_version = "2.9"
nearest_model_name = "resnet50"
estimator.register(
content_types=content_types,
response_types=response_types,
inference_instances=inference_instances,
transform_instances=transform_instances,
model_package_name=model_package_name,
sample_payload_url=sample_payload_url,
task=task,
image_uri=inference_image,
framework=framework,
framework_version=framework_version,
nearest_model_name=nearest_model_name,
)
sagemaker_session.create_model.assert_not_called()
expected_create_model_package_request = {
"containers": [
{
"Image": inference_image,
"ModelDataUrl": estimator.model_data,
}
],
"content_types": content_types,
"response_types": response_types,
"inference_instances": inference_instances,
"transform_instances": transform_instances,
"model_package_name": model_package_name,
"marketplace_cert": False,
"sample_payload_url": sample_payload_url,
"task": task,
}
sagemaker_session.create_model_package_from_containers.assert_called_with(
**expected_create_model_package_request
)
def test_register_under_pipeline_session(pipeline_session):
estimator = Estimator(
IMAGE_URI,
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
output_path=OUTPUT_PATH,
sagemaker_session=pipeline_session,
)
model_package_name = "test-estimator-register-model"
content_types = ["application/json"]
response_types = ["application/json"]
inference_instances = ["ml.m4.xlarge"]
transform_instances = ["ml.m4.xlarget"]
with pytest.raises(TypeError) as error:
estimator.register(
content_types=content_types,
response_types=response_types,
inference_instances=inference_instances,
transform_instances=transform_instances,
model_package_name=model_package_name,
)
assert "estimator.register does not support PipelineSession" in str(error.value)
@patch("sagemaker.estimator.LocalSession")
@patch("sagemaker.estimator.Session")
def test_local_mode(session_class, local_session_class):
local_session = Mock(spec=sagemaker.local.LocalSession)
local_session.local_mode = True
session = Mock()
session.local_mode = False
local_session_class.return_value = local_session
session_class.return_value = session
e = Estimator(IMAGE_URI, ROLE, INSTANCE_COUNT, "local")
assert e.sagemaker_session.local_mode is True
e2 = Estimator(IMAGE_URI, ROLE, INSTANCE_COUNT, "local_gpu")
assert e2.sagemaker_session.local_mode is True
e3 = Estimator(IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE)
assert e3.sagemaker_session.local_mode is False
@patch("sagemaker.estimator.LocalSession")
def test_distributed_gpu_local_mode(LocalSession):
with pytest.raises(RuntimeError):
Estimator(IMAGE_URI, ROLE, 3, "local_gpu", output_path=OUTPUT_PATH)
@patch("sagemaker.estimator.LocalSession")
def test_local_mode_file_output_path(local_session_class):
local_session = Mock(spec=sagemaker.local.LocalSession)
local_session.local_mode = True
local_session_class.return_value = local_session
e = Estimator(IMAGE_URI, ROLE, INSTANCE_COUNT, "local", output_path="file:///tmp/model/")
assert e.output_path == "file:///tmp/model/"
@patch("sagemaker.estimator.Session")
def test_file_output_path_not_supported_outside_local_mode(session_class):
session = Mock()
session.local_mode = False
session_class.return_value = session
with pytest.raises(RuntimeError):
Estimator(IMAGE_URI, ROLE, INSTANCE_COUNT, INSTANCE_TYPE, output_path="file:///tmp/model")
def test_prepare_init_params_from_job_description_with_image_training_job():
init_params = EstimatorBase._prepare_init_params_from_job_description(
job_details=RETURNED_JOB_DESCRIPTION
)
assert init_params["role"] == "arn:aws:iam::366:role/SageMakerRole"
assert init_params["instance_count"] == 1
assert init_params["image_uri"] == "1.dkr.ecr.us-west-2.amazonaws.com/sagemaker-other:1.0.4"
def test_prepare_init_params_from_job_description_with_algorithm_training_job():
algorithm_job_description = RETURNED_JOB_DESCRIPTION.copy()
algorithm_job_description["AlgorithmSpecification"] = {
"TrainingInputMode": "File",
"AlgorithmName": "arn:aws:sagemaker:us-east-2:1234:algorithm/scikit-decision-trees",
"TrainingImage": "",
}
init_params = EstimatorBase._prepare_init_params_from_job_description(
job_details=algorithm_job_description
)
assert init_params["role"] == "arn:aws:iam::366:role/SageMakerRole"
assert init_params["instance_count"] == 1
assert (
init_params["algorithm_arn"]
== "arn:aws:sagemaker:us-east-2:1234:algorithm/scikit-decision-trees"
)
def test_prepare_init_params_from_job_description_with_spot_training():
job_description = RETURNED_JOB_DESCRIPTION.copy()
job_description["EnableManagedSpotTraining"] = True
job_description["StoppingCondition"] = {
"MaxRuntimeInSeconds": 86400,
"MaxWaitTimeInSeconds": 87000,
}
init_params = EstimatorBase._prepare_init_params_from_job_description(
job_details=job_description
)
assert init_params["role"] == "arn:aws:iam::366:role/SageMakerRole"
assert init_params["instance_count"] == 1
assert init_params["use_spot_instances"]
assert init_params["max_run"] == 86400
assert init_params["max_wait"] == 87000
def test_prepare_init_params_from_job_description_with_retry_strategy():
job_description = RETURNED_JOB_DESCRIPTION.copy()
job_description["RetryStrategy"] = {"MaximumRetryAttempts": 2}
job_description["StoppingCondition"] = {
"MaxRuntimeInSeconds": 86400,
"MaxWaitTimeInSeconds": 87000,
}
init_params = EstimatorBase._prepare_init_params_from_job_description(
job_details=job_description
)
assert init_params["role"] == "arn:aws:iam::366:role/SageMakerRole"
assert init_params["instance_count"] == 1
assert init_params["max_run"] == 86400
assert init_params["max_wait"] == 87000
assert init_params["max_retry_attempts"] == 2
def test_prepare_init_params_from_job_description_with_invalid_training_job():
invalid_job_description = RETURNED_JOB_DESCRIPTION.copy()
invalid_job_description["AlgorithmSpecification"] = {"TrainingInputMode": "File"}
with pytest.raises(RuntimeError) as error:
EstimatorBase._prepare_init_params_from_job_description(job_details=invalid_job_description)
assert "Invalid AlgorithmSpecification" in str(error)
def test_prepare_for_training_with_base_name(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_count=1,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
estimator._prepare_for_training()
assert "base_job_name" in estimator._current_job_name
def test_prepare_for_training_with_name_based_on_image(sagemaker_session):
estimator = Estimator(
image_uri="some-image",
role="some_image",
instance_count=1,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
)
estimator._prepare_for_training()
assert "some-image" in estimator._current_job_name
@patch("sagemaker.algorithm.AlgorithmEstimator.validate_train_spec", Mock())
@patch("sagemaker.algorithm.AlgorithmEstimator._parse_hyperparameters", Mock(return_value={}))
def test_prepare_for_training_with_name_based_on_algorithm(sagemaker_session):
estimator = AlgorithmEstimator(
algorithm_arn="arn:aws:sagemaker:us-west-2:1234:algorithm/scikit-decision-trees-1542410022",
role="some_image",
instance_count=1,
instance_type="ml.m4.xlarge",
sagemaker_session=sagemaker_session,
)
estimator._prepare_for_training()
assert "scikit-decision-trees-1542410022" in estimator._current_job_name
@patch(
"sagemaker.estimator.Estimator.fit",
Mock(
side_effect=ClientError(
error_response={
"Error": {
"Code": 403,
"Message": '"EnableInterContainerTrafficEncryption" and '
'"VpcConfig" must be provided together',
}
},
operation_name="Unit Test",
)
),
)
def test_encryption_flag_in_non_vpc_mode_invalid(sagemaker_session):
with pytest.raises(ClientError) as error:
estimator = Estimator(
image_uri="some-image",
role="SageMakerRole",
instance_count=1,
instance_type="ml.c4.xlarge",
sagemaker_session=sagemaker_session,
base_job_name="test-non-vpc-encryption",
encrypt_inter_container_traffic=True,
)
estimator.fit()
assert (
'"EnableInterContainerTrafficEncryption" and "VpcConfig" must be provided together'
in str(error)
)
def test_estimator_local_mode_error(sagemaker_session):
# When using instance local with a session which is not LocalSession we should error out
with pytest.raises(RuntimeError):
Estimator(
image_uri="some-image",
role="some_image",
instance_count=1,
instance_type="local",
sagemaker_session=sagemaker_session,
base_job_name="base_job_name",
)
def test_estimator_local_mode_ok(sagemaker_local_session):
# When using instance local with a session which is not LocalSession we should error out
Estimator(
image_uri="some-image",
role="some_image",
instance_count=1,
instance_type="local",
sagemaker_session=sagemaker_local_session,
base_job_name="base_job_name",
)
def test_framework_distribution_configuration(sagemaker_session):
framework = DummyFramework(
entry_point="script",
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
actual_ps = framework._distribution_configuration(distribution=DISTRIBUTION_PS_ENABLED)
expected_ps = {"sagemaker_parameter_server_enabled": True}
assert actual_ps == expected_ps
actual_mpi = framework._distribution_configuration(distribution=DISTRIBUTION_MPI_ENABLED)
expected_mpi = {
"sagemaker_mpi_enabled": True,
"sagemaker_mpi_num_of_processes_per_host": 2,
"sagemaker_mpi_custom_mpi_options": "options",
}
assert actual_mpi == expected_mpi
actual_ddp = framework._distribution_configuration(distribution=DISTRIBUTION_SM_DDP_ENABLED)
expected_ddp = {
"sagemaker_distributed_dataparallel_enabled": True,
"sagemaker_distributed_dataparallel_custom_mpi_options": "options",
"sagemaker_instance_type": INSTANCE_TYPE,
}
assert actual_ddp == expected_ddp
def test_image_name_map(sagemaker_session):
e = DummyFramework(
"my_script.py",
image_name=IMAGE_URI,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
)
assert e.image_uri == IMAGE_URI
@patch("sagemaker.git_utils.git_clone_repo")
def test_git_support_with_branch_and_commit_succeed_estimator_class(
git_clone_repo, sagemaker_session
):
git_clone_repo.side_effect = lambda gitconfig, entrypoint, source_dir=None, dependencies=None: {
"entry_point": "/tmp/repo_dir/entry_point",
"source_dir": None,
"dependencies": None,
}
git_config = {"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT}
entry_point = "entry_point"
fw = Estimator(
entry_point=entry_point,
git_config=git_config,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
image_uri=IMAGE_URI,
)
fw.fit()
git_clone_repo.assert_called_once_with(git_config, entry_point, None, [])
@patch("sagemaker.estimator.Estimator._stage_user_code_in_s3")
def test_script_mode_estimator(patched_stage_user_code, sagemaker_session):
patched_stage_user_code.return_value = UploadedCode(
s3_prefix="s3://bucket/key", script_name="script_name"
)
script_uri = "s3://codebucket/someprefix/sourcedir.tar.gz"
image_uri = "763104351884.dkr.ecr.us-west-2.amazonaws.com/pytorch-training:1.9.0-gpu-py38"
model_uri = "s3://someprefix2/models/model.tar.gz"
t = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
sagemaker_session=sagemaker_session,
instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
source_dir=script_uri,
image_uri=image_uri,
model_uri=model_uri,
)
t.fit("s3://bucket/mydata")
patched_stage_user_code.assert_called_once()
sagemaker_session.train.assert_called_once()
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
def test_script_mode_estimator_same_calls_as_framework(
patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
script_uri = "s3://codebucket/someprefix/sourcedir.tar.gz"
instance_type = "ml.p2.xlarge"
instance_count = 1
model_uri = "s3://someprefix2/models/model.tar.gz"
training_data_uri = "s3://bucket/mydata"
hyperparameters = {
"int_hyperparam": 1,
"string_hyperparam": "hello",
"stringified_numeric_hyperparam": "44",
"float_hyperparam": 1.234,
}
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=script_uri,
image_uri=IMAGE_URI,
model_uri=model_uri,
dependencies=[],
debugger_hook_config={},
hyperparameters=deepcopy(hyperparameters),
)
generic_estimator.fit(training_data_uri)
generic_estimator_tar_and_upload_dir_args = patched_tar_and_upload_dir.call_args_list
generic_estimator_train_args = sagemaker_session.train.call_args_list
patched_tar_and_upload_dir.reset_mock()
sagemaker_session.train.reset_mock()
framework_estimator = DummyFramework(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
source_dir=script_uri,
instance_count=instance_count,
instance_type=instance_type,
sagemaker_session=sagemaker_session,
model_uri=model_uri,
dependencies=[],
debugger_hook_config={},
hyperparameters=deepcopy(hyperparameters),
)
framework_estimator.fit(training_data_uri)
assert len(generic_estimator_tar_and_upload_dir_args) == 1
assert len(generic_estimator_train_args) == 1
assert generic_estimator_tar_and_upload_dir_args == patched_tar_and_upload_dir.call_args_list
assert generic_estimator_train_args == sagemaker_session.train.call_args_list
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
def test_script_mode_estimator_tags_jumpstart_estimators_and_models(
patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
instance_type = "ml.p2.xlarge"
instance_count = 1
training_data_uri = "s3://bucket/mydata"
jumpstart_source_dir = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/source.tar.gz"
jumpstart_source_dir_2 = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[1]}/source_dirs/source.tar.gz"
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=jumpstart_source_dir,
image_uri=IMAGE_URI,
model_uri=jumpstart_source_dir_2,
tags=[{"Key": "some", "Value": "tag"}],
)
generic_estimator.fit(training_data_uri)
assert [
{"Key": "some", "Value": "tag"},
{
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_source_dir_2,
},
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
] == sagemaker_session.train.call_args_list[0][1]["tags"]
sagemaker_session.reset_mock()
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
inference_jumpstart_source_dir = (
f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/inference/source.tar.gz"
)
generic_estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
image_uri=IMAGE_URI,
source_dir=inference_jumpstart_source_dir,
entry_point="inference.py",
role=ROLE,
tags=[{"Key": "deploys", "Value": "tag"}],
)
assert sagemaker_session.create_model.call_args_list[0][1]["tags"] == [
{"Key": "deploys", "Value": "tag"},
{
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_source_dir_2,
},
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": inference_jumpstart_source_dir,
},
]
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0][1]["tags"] == [
{"Key": "deploys", "Value": "tag"},
{
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_source_dir_2,
},
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": inference_jumpstart_source_dir,
},
]
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
def test_script_mode_estimator_tags_jumpstart_models(
patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
instance_type = "ml.p2.xlarge"
instance_count = 1
training_data_uri = "s3://bucket/mydata"
jumpstart_source_dir = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/source.tar.gz"
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=jumpstart_source_dir,
image_uri=IMAGE_URI,
model_uri=MODEL_DATA,
)
generic_estimator.fit(training_data_uri)
assert [
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
] == sagemaker_session.train.call_args_list[0][1]["tags"]
sagemaker_session.reset_mock()
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
inference_source_dir = "s3://dsfsdfsd/sdfsdfs/sdfsd"
generic_estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
image_uri=IMAGE_URI,
source_dir=inference_source_dir,
entry_point="inference.py",
role=ROLE,
)
assert sagemaker_session.create_model.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
]
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.TRAINING_SCRIPT_URI.value,
"Value": jumpstart_source_dir,
},
]
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
def test_script_mode_estimator_tags_jumpstart_models_with_no_estimator_js_tags(
patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
instance_type = "ml.p2.xlarge"
instance_count = 1
training_data_uri = "s3://bucket/mydata"
source_dir = "s3://dsfsdfsd/sdfsdfs/sdfsd"
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=source_dir,
image_uri=IMAGE_URI,
model_uri=MODEL_DATA,
)
generic_estimator.fit(training_data_uri)
assert None is sagemaker_session.train.call_args_list[0][1]["tags"]
sagemaker_session.reset_mock()
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
inference_jumpstart_source_dir = (
f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/inference/source.tar.gz"
)
generic_estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
image_uri=IMAGE_URI,
source_dir=inference_jumpstart_source_dir,
entry_point="inference.py",
role=ROLE,
)
assert sagemaker_session.create_model.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": inference_jumpstart_source_dir,
},
]
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": inference_jumpstart_source_dir,
},
]
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
@patch("sagemaker.utils.repack_model")
def test_all_framework_estimators_add_jumpstart_tags(
patched_repack_model, patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
sagemaker_session.boto_region_name = REGION
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
framework_estimator_classes_to_kwargs = {
PyTorch: {
"framework_version": "1.5.0",
"py_version": "py3",
"instance_type": "ml.p2.xlarge",
},
TensorFlow: {
"framework_version": "2.3",
"py_version": "py37",
"instance_type": "ml.p2.xlarge",
},
HuggingFace: {
"pytorch_version": "1.7.1",
"py_version": "py36",
"transformers_version": "4.6.1",
"instance_type": "ml.p2.xlarge",
},
MXNet: {"framework_version": "1.7.0", "py_version": "py3", "instance_type": "ml.p2.xlarge"},
SKLearn: {"framework_version": "0.23-1", "instance_type": "ml.m2.xlarge"},
XGBoost: {"framework_version": "1.3-1", "instance_type": "ml.m2.xlarge"},
}
jumpstart_model_uri = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/model_dirs/model.tar.gz"
jumpstart_model_uri_2 = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[1]}/model_dirs/model.tar.gz"
for framework_estimator_class, kwargs in framework_estimator_classes_to_kwargs.items():
estimator = framework_estimator_class(
entry_point=ENTRY_POINT,
role=ROLE,
sagemaker_session=sagemaker_session,
model_uri=jumpstart_model_uri,
instance_count=INSTANCE_COUNT,
**kwargs,
)
estimator.fit()
assert {
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_model_uri,
} in sagemaker_session.train.call_args_list[0][1]["tags"]
estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=kwargs["instance_type"],
image_uri=IMAGE_URI,
source_dir=jumpstart_model_uri_2,
entry_point="inference.py",
role=ROLE,
)
assert sagemaker_session.create_model.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_model_uri,
},
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": jumpstart_model_uri_2,
},
]
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0][1]["tags"] == [
{
"Key": JumpStartTag.TRAINING_MODEL_URI.value,
"Value": jumpstart_model_uri,
},
{
"Key": JumpStartTag.INFERENCE_SCRIPT_URI.value,
"Value": jumpstart_model_uri_2,
},
]
sagemaker_session.train.reset_mock()
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
def test_script_mode_estimator_uses_jumpstart_base_name_with_js_models(
patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
instance_type = "ml.p2.xlarge"
instance_count = 1
training_data_uri = "s3://bucket/mydata"
source_dir = "s3://dsfsdfsd/sdfsdfs/sdfsd"
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=source_dir,
image_uri=IMAGE_URI,
model_uri=MODEL_DATA,
)
generic_estimator.fit(training_data_uri)
assert not sagemaker_session.train.call_args_list[0][1]["job_name"].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
sagemaker_session.reset_mock()
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
inference_jumpstart_source_dir = (
f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/inference/source.tar.gz"
)
generic_estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=INSTANCE_TYPE,
image_uri=IMAGE_URI,
source_dir=inference_jumpstart_source_dir,
entry_point="inference.py",
role=ROLE,
)
assert sagemaker_session.create_model.call_args_list[0][1]["name"].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
@patch("sagemaker.utils.repack_model")
def test_all_framework_estimators_add_jumpstart_base_name(
patched_repack_model, patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
sagemaker_session.boto_region_name = REGION
sagemaker_session.sagemaker_client.describe_training_job.return_value = {
"ModelArtifacts": {"S3ModelArtifacts": "some-uri"}
}
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
framework_estimator_classes_to_kwargs = {
PyTorch: {
"framework_version": "1.5.0",
"py_version": "py3",
"instance_type": "ml.p2.xlarge",
},
TensorFlow: {
"framework_version": "2.3",
"py_version": "py37",
"instance_type": "ml.p2.xlarge",
},
HuggingFace: {
"pytorch_version": "1.7.1",
"py_version": "py36",
"transformers_version": "4.6.1",
"instance_type": "ml.p2.xlarge",
},
MXNet: {"framework_version": "1.7.0", "py_version": "py3", "instance_type": "ml.p2.xlarge"},
SKLearn: {"framework_version": "0.23-1", "instance_type": "ml.m2.xlarge"},
XGBoost: {"framework_version": "1.3-1", "instance_type": "ml.m2.xlarge"},
}
jumpstart_model_uri = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/model_dirs/model.tar.gz"
jumpstart_model_uri_2 = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[1]}/model_dirs/model.tar.gz"
for framework_estimator_class, kwargs in framework_estimator_classes_to_kwargs.items():
estimator = framework_estimator_class(
entry_point=ENTRY_POINT,
role=ROLE,
sagemaker_session=sagemaker_session,
model_uri=jumpstart_model_uri,
instance_count=INSTANCE_COUNT,
**kwargs,
)
estimator.fit()
assert sagemaker_session.train.call_args_list[0][1]["job_name"].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
estimator.deploy(
initial_instance_count=INSTANCE_COUNT,
instance_type=kwargs["instance_type"],
image_uri=IMAGE_URI,
source_dir=jumpstart_model_uri_2,
entry_point="inference.py",
role=ROLE,
)
assert sagemaker_session.create_model.call_args_list[0][1]["name"].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
assert sagemaker_session.endpoint_from_production_variants.call_args_list[0].startswith(
JUMPSTART_RESOURCE_BASE_NAME
)
sagemaker_session.endpoint_from_production_variants.reset_mock()
sagemaker_session.create_model.reset_mock()
sagemaker_session.train.reset_mock()
def test_insert_invalid_source_code_args():
with pytest.raises(TypeError) as err:
Estimator(
image_uri="IMAGE_URI",
role=ROLE,
entry_point=ParameterString(name="EntryPoint"),
instance_type="ml.m5.xlarge",
instance_count=1,
enable_network_isolation=True,
)
assert (
"entry_point, source_dir should not be pipeline variables "
"when enable_network_isolation is a pipeline variable or it is set to True."
) in str(err.value)
with pytest.raises(TypeError) as err:
Estimator(
image_uri="IMAGE_URI",
role=ROLE,
entry_point="dummy.py",
source_dir=ParameterString(name="SourceDir"),
instance_type="ml.m5.xlarge",
instance_count=1,
enable_network_isolation=ParameterBoolean(name="EnableNetworkIsolation"),
)
assert (
"entry_point, source_dir should not be pipeline variables "
"when enable_network_isolation is a pipeline variable or it is set to True."
) in str(err.value)
with pytest.raises(TypeError) as err:
Estimator(
image_uri=IMAGE_URI,
role=ROLE,
git_config={"repo": GIT_REPO, "branch": BRANCH, "commit": COMMIT},
source_dir=ParameterString(name="SourceDir"),
entry_point=ParameterString(name="EntryPoint"),
instance_type="ml.m5.xlarge",
instance_count=1,
)
assert (
"entry_point, source_dir should not be pipeline variables when git_config is given"
in str(err.value)
)
with pytest.raises(TypeError) as err:
Estimator(
image_uri=IMAGE_URI,
role=ROLE,
entry_point=ParameterString(name="EntryPoint"),
instance_type="ml.m5.xlarge",
instance_count=1,
)
assert "The entry_point should not be a pipeline variable when source_dir is missing" in str(
err.value
)
with pytest.raises(TypeError) as err:
Estimator(
image_uri="IMAGE_URI",
role=ROLE,
entry_point=ParameterString(name="EntryPoint"),
source_dir="file://my-file/",
instance_type="ml.m5.xlarge",
instance_count=1,
)
assert (
"The entry_point should not be a pipeline variable " "when source_dir is a local path"
) in str(err.value)
@patch("time.time", return_value=TIME)
@patch("sagemaker.estimator.tar_and_upload_dir")
@patch("sagemaker.model.Model._upload_code")
def test_script_mode_estimator_escapes_hyperparameters_as_json(
patched_upload_code, patched_tar_and_upload_dir, sagemaker_session
):
patched_tar_and_upload_dir.return_value = UploadedCode(
s3_prefix="s3://%s/%s" % ("bucket", "key"), script_name="script_name"
)
sagemaker_session.boto_region_name = REGION
instance_type = "ml.p2.xlarge"
instance_count = 1
training_data_uri = "s3://bucket/mydata"
jumpstart_source_dir = f"s3://{list(JUMPSTART_BUCKET_NAME_SET)[0]}/source_dirs/source.tar.gz"
hyperparameters = {
"int_hyperparam": 1,
"string_hyperparam": "hello",
"stringified_numeric_hyperparam": "44",
"float_hyperparam": 1.234,
}
generic_estimator = Estimator(
entry_point=SCRIPT_PATH,
role=ROLE,
region=REGION,
sagemaker_session=sagemaker_session,
instance_count=instance_count,
instance_type=instance_type,
source_dir=jumpstart_source_dir,
image_uri=IMAGE_URI,
model_uri=MODEL_DATA,
hyperparameters=hyperparameters,
)
generic_estimator.fit(training_data_uri)
formatted_hyperparams = EstimatorBase._json_encode_hyperparameters(hyperparameters)
assert (
set(formatted_hyperparams.items())
- set(sagemaker_session.train.call_args_list[0][1]["hyperparameters"].items())
== set()
)