hc99's picture
Add files using upload-large-folder tool
476455e verified
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"). You
# may not use this file except in compliance with the License. A copy of
# the License is located at
#
# http://aws.amazon.com/apache2.0/
#
# or in the "license" file accompanying this file. This file is
# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
# ANY KIND, either express or implied. See the License for the specific
# language governing permissions and limitations under the License.
from __future__ import absolute_import
import pytest
from mock import Mock, patch
from sagemaker import image_uris
from sagemaker.amazon.linear_learner import LinearLearner, LinearLearnerPredictor
from sagemaker.amazon.amazon_estimator import RecordSet
ROLE = "myrole"
INSTANCE_COUNT = 1
INSTANCE_TYPE = "ml.c4.xlarge"
PREDICTOR_TYPE = "binary_classifier"
COMMON_TRAIN_ARGS = {
"role": ROLE,
"instance_count": INSTANCE_COUNT,
"instance_type": INSTANCE_TYPE,
}
ALL_REQ_ARGS = dict({"predictor_type": PREDICTOR_TYPE}, **COMMON_TRAIN_ARGS)
REGION = "us-west-2"
BUCKET_NAME = "Some-Bucket"
DESCRIBE_TRAINING_JOB_RESULT = {"ModelArtifacts": {"S3ModelArtifacts": "s3://bucket/model.tar.gz"}}
ENDPOINT_DESC = {"EndpointConfigName": "test-endpoint"}
ENDPOINT_CONFIG_DESC = {"ProductionVariants": [{"ModelName": "model-1"}, {"ModelName": "model-2"}]}
@pytest.fixture()
def sagemaker_session():
boto_mock = Mock(name="boto_session", region_name=REGION)
sms = Mock(
name="sagemaker_session",
boto_session=boto_mock,
region_name=REGION,
config=None,
local_mode=False,
s3_client=None,
s3_resource=None,
)
sms.boto_region_name = REGION
sms.default_bucket = Mock(name="default_bucket", return_value=BUCKET_NAME)
sms.sagemaker_client.describe_training_job = Mock(
name="describe_training_job", return_value=DESCRIBE_TRAINING_JOB_RESULT
)
sms.sagemaker_client.describe_endpoint = Mock(return_value=ENDPOINT_DESC)
sms.sagemaker_client.describe_endpoint_config = Mock(return_value=ENDPOINT_CONFIG_DESC)
return sms
def test_init_required_positional(sagemaker_session):
lr = LinearLearner(
ROLE,
INSTANCE_COUNT,
INSTANCE_TYPE,
PREDICTOR_TYPE,
sagemaker_session=sagemaker_session,
)
assert lr.role == ROLE
assert lr.instance_count == INSTANCE_COUNT
assert lr.instance_type == INSTANCE_TYPE
assert lr.predictor_type == PREDICTOR_TYPE
def test_init_required_named(sagemaker_session):
lr = LinearLearner(sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
assert lr.role == ALL_REQ_ARGS["role"]
assert lr.instance_count == ALL_REQ_ARGS["instance_count"]
assert lr.instance_type == ALL_REQ_ARGS["instance_type"]
assert lr.predictor_type == ALL_REQ_ARGS["predictor_type"]
def test_all_hyperparameters(sagemaker_session):
lr = LinearLearner(
sagemaker_session=sagemaker_session,
binary_classifier_model_selection_criteria="accuracy",
target_recall=0.5,
target_precision=0.6,
positive_example_weight_mult=0.1,
epochs=1,
use_bias=True,
num_models=5,
num_calibration_samples=6,
init_method="uniform",
init_scale=0.1,
init_sigma=0.001,
init_bias=0,
optimizer="sgd",
loss="logistic",
wd=0.4,
l1=0.04,
momentum=0.1,
learning_rate=0.001,
beta_1=0.2,
beta_2=0.03,
bias_lr_mult=5.5,
bias_wd_mult=6.6,
use_lr_scheduler=False,
lr_scheduler_step=2,
lr_scheduler_factor=0.03,
lr_scheduler_minimum_lr=0.001,
normalize_data=False,
normalize_label=True,
unbias_data=True,
unbias_label=False,
num_point_for_scaler=3,
margin=1.0,
quantile=0.5,
loss_insensitivity=0.1,
huber_delta=0.1,
early_stopping_patience=3,
early_stopping_tolerance=0.001,
num_classes=1,
accuracy_top_k=3,
f_beta=1.0,
balance_multiclass_weights=False,
**ALL_REQ_ARGS,
)
assert lr.hyperparameters() == dict(
predictor_type="binary_classifier",
binary_classifier_model_selection_criteria="accuracy",
target_recall="0.5",
target_precision="0.6",
positive_example_weight_mult="0.1",
epochs="1",
use_bias="True",
num_models="5",
num_calibration_samples="6",
init_method="uniform",
init_scale="0.1",
init_sigma="0.001",
init_bias="0.0",
optimizer="sgd",
loss="logistic",
wd="0.4",
l1="0.04",
momentum="0.1",
learning_rate="0.001",
beta_1="0.2",
beta_2="0.03",
bias_lr_mult="5.5",
bias_wd_mult="6.6",
use_lr_scheduler="False",
lr_scheduler_step="2",
lr_scheduler_factor="0.03",
lr_scheduler_minimum_lr="0.001",
normalize_data="False",
normalize_label="True",
unbias_data="True",
unbias_label="False",
num_point_for_scaler="3",
margin="1.0",
quantile="0.5",
loss_insensitivity="0.1",
huber_delta="0.1",
early_stopping_patience="3",
early_stopping_tolerance="0.001",
num_classes="1",
accuracy_top_k="3",
f_beta="1.0",
balance_multiclass_weights="False",
)
def test_image(sagemaker_session):
lr = LinearLearner(sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
assert image_uris.retrieve("linear-learner", REGION) == lr.training_image_uri()
@pytest.mark.parametrize("required_hyper_parameters, value", [("predictor_type", 0)])
def test_required_hyper_parameters_type(sagemaker_session, required_hyper_parameters, value):
with pytest.raises(ValueError):
test_params = ALL_REQ_ARGS.copy()
test_params[required_hyper_parameters] = value
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
@pytest.mark.parametrize("required_hyper_parameters, value", [("predictor_type", "string")])
def test_required_hyper_parameters_value(sagemaker_session, required_hyper_parameters, value):
with pytest.raises(ValueError):
test_params = ALL_REQ_ARGS.copy()
test_params[required_hyper_parameters] = value
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
def test_num_classes_is_required_for_multiclass_classifier(sagemaker_session):
with pytest.raises(ValueError) as excinfo:
test_params = ALL_REQ_ARGS.copy()
test_params["predictor_type"] = "multiclass_classifier"
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
assert (
"For predictor_type 'multiclass_classifier', 'num_classes' should be set to a value greater than 2."
in str(excinfo.value)
)
def test_num_classes_can_be_string_for_multiclass_classifier(sagemaker_session):
test_params = ALL_REQ_ARGS.copy()
test_params["predictor_type"] = "multiclass_classifier"
test_params["num_classes"] = "3"
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
@pytest.mark.parametrize("iterable_hyper_parameters, value", [("epochs", [0])])
def test_iterable_hyper_parameters_type(sagemaker_session, iterable_hyper_parameters, value):
with pytest.raises(TypeError):
test_params = ALL_REQ_ARGS.copy()
test_params.update({iterable_hyper_parameters: value})
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
@pytest.mark.parametrize(
"optional_hyper_parameters, value",
[
("binary_classifier_model_selection_criteria", 0),
("target_recall", "string"),
("target_precision", "string"),
("epochs", "string"),
("num_models", "string"),
("num_calibration_samples", "string"),
("init_method", 0),
("init_scale", "string"),
("init_sigma", "string"),
("init_bias", "string"),
("optimizer", 0),
("loss", 0),
("wd", "string"),
("l1", "string"),
("momentum", "string"),
("learning_rate", "string"),
("beta_1", "string"),
("beta_2", "string"),
("bias_lr_mult", "string"),
("bias_wd_mult", "string"),
("lr_scheduler_step", "string"),
("lr_scheduler_factor", "string"),
("lr_scheduler_minimum_lr", "string"),
("num_point_for_scaler", "string"),
("margin", "string"),
("quantile", "string"),
("loss_insensitivity", "string"),
("huber_delta", "string"),
("early_stopping_patience", "string"),
("early_stopping_tolerance", "string"),
("num_classes", "string"),
("accuracy_top_k", "string"),
("f_beta", "string"),
],
)
def test_optional_hyper_parameters_type(sagemaker_session, optional_hyper_parameters, value):
with pytest.raises(ValueError):
test_params = ALL_REQ_ARGS.copy()
test_params.update({optional_hyper_parameters: value})
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
@pytest.mark.parametrize(
"optional_hyper_parameters, value",
[
("binary_classifier_model_selection_criteria", "string"),
("target_recall", 0),
("target_recall", 1),
("target_precision", 0),
("target_precision", 1),
("epochs", 0),
("num_models", 0),
("num_calibration_samples", 0),
("init_method", "string"),
("init_scale", 0),
("init_sigma", 0),
("optimizer", "string"),
("loss", "string"),
("wd", -1),
("l1", -1),
("momentum", 1),
("learning_rate", 0),
("beta_1", 1),
("beta_2", 1),
("bias_lr_mult", 0),
("bias_wd_mult", -1),
("lr_scheduler_step", 0),
("lr_scheduler_factor", 0),
("lr_scheduler_factor", 1),
("lr_scheduler_minimum_lr", 0),
("num_point_for_scaler", 0),
("margin", -1),
("quantile", 0),
("quantile", 1),
("loss_insensitivity", 0),
("huber_delta", -1),
("early_stopping_patience", 0),
("early_stopping_tolerance", 0),
("num_classes", 0),
("accuracy_top_k", 0),
("f_beta", -1.0),
],
)
def test_optional_hyper_parameters_value(sagemaker_session, optional_hyper_parameters, value):
with pytest.raises(ValueError):
test_params = ALL_REQ_ARGS.copy()
test_params.update({optional_hyper_parameters: value})
LinearLearner(sagemaker_session=sagemaker_session, **test_params)
PREFIX = "prefix"
FEATURE_DIM = 10
DEFAULT_MINI_BATCH_SIZE = 1000
def test_prepare_for_training_calculate_batch_size_1(sagemaker_session):
lr = LinearLearner(base_job_name="lr", sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=1,
feature_dim=FEATURE_DIM,
channel="train",
)
lr._prepare_for_training(data)
assert lr.mini_batch_size == 1
def test_prepare_for_training_calculate_batch_size_2(sagemaker_session):
lr = LinearLearner(base_job_name="lr", sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=10000,
feature_dim=FEATURE_DIM,
channel="train",
)
lr._prepare_for_training(data)
assert lr.mini_batch_size == DEFAULT_MINI_BATCH_SIZE
def test_prepare_for_training_multiple_channel(sagemaker_session):
lr = LinearLearner(base_job_name="lr", sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=10000,
feature_dim=FEATURE_DIM,
channel="train",
)
lr._prepare_for_training([data, data])
assert lr.mini_batch_size == DEFAULT_MINI_BATCH_SIZE
def test_prepare_for_training_multiple_channel_no_train(sagemaker_session):
lr = LinearLearner(base_job_name="lr", sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=10000,
feature_dim=FEATURE_DIM,
channel="mock",
)
with pytest.raises(ValueError) as ex:
lr._prepare_for_training([data, data])
assert "Must provide train channel." in str(ex)
@patch("sagemaker.amazon.amazon_estimator.AmazonAlgorithmEstimatorBase.fit")
def test_call_fit_pass_batch_size(base_fit, sagemaker_session):
lr = LinearLearner(base_job_name="lr", sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=10000,
feature_dim=FEATURE_DIM,
channel="train",
)
lr.fit(data, 10)
base_fit.assert_called_once()
assert len(base_fit.call_args[0]) == 2
assert base_fit.call_args[0][0] == data
assert base_fit.call_args[0][1] == 10
def test_model_image(sagemaker_session):
lr = LinearLearner(sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=1,
feature_dim=FEATURE_DIM,
channel="train",
)
lr.fit(data)
model = lr.create_model()
assert image_uris.retrieve("linear-learner", REGION) == model.image_uri
def test_predictor_type(sagemaker_session):
lr = LinearLearner(sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=1,
feature_dim=FEATURE_DIM,
channel="train",
)
lr.fit(data)
model = lr.create_model()
predictor = model.deploy(1, INSTANCE_TYPE)
assert isinstance(predictor, LinearLearnerPredictor)
def test_predictor_custom_serialization(sagemaker_session):
lr = LinearLearner(sagemaker_session=sagemaker_session, **ALL_REQ_ARGS)
data = RecordSet(
"s3://{}/{}".format(BUCKET_NAME, PREFIX),
num_records=1,
feature_dim=FEATURE_DIM,
channel="train",
)
lr.fit(data)
model = lr.create_model()
custom_serializer = Mock()
custom_deserializer = Mock()
predictor = model.deploy(
1,
INSTANCE_TYPE,
serializer=custom_serializer,
deserializer=custom_deserializer,
)
assert isinstance(predictor, LinearLearnerPredictor)
assert predictor.serializer is custom_serializer
assert predictor.deserializer is custom_deserializer