hc99's picture
Add files using upload-large-folder tool
476455e verified
# Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"). You
# may not use this file except in compliance with the License. A copy of
# the License is located at
#
# http://aws.amazon.com/apache2.0/
#
# or in the "license" file accompanying this file. This file is
# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF
# ANY KIND, either express or implied. See the License for the specific
# language governing permissions and limitations under the License.
from __future__ import absolute_import
import datetime
import io
import logging
import os
import pytest
import six
from botocore.exceptions import ClientError
from mock import ANY, MagicMock, Mock, patch, call, mock_open
import sagemaker
from sagemaker import TrainingInput, Session, get_execution_role
from sagemaker.async_inference import AsyncInferenceConfig
from sagemaker.session import (
_tuning_job_status,
_transform_job_status,
_train_done,
NOTEBOOK_METADATA_FILE,
)
from sagemaker.tuner import WarmStartConfig, WarmStartTypes
from sagemaker.inputs import BatchDataCaptureConfig
STATIC_HPs = {"feature_dim": "784"}
SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}]
ENV_INPUT = {"env_key1": "env_val1", "env_key2": "env_val2", "env_key3": "env_val3"}
REGION = "us-west-2"
STS_ENDPOINT = "sts.us-west-2.amazonaws.com"
@pytest.fixture()
def boto_session():
boto_mock = Mock(name="boto_session", region_name=REGION)
client_mock = Mock()
client_mock._client_config.user_agent = (
"Boto3/1.9.69 Python/3.6.5 Linux/4.14.77-70.82.amzn1.x86_64 Botocore/1.12.69 Resource"
)
boto_mock.client.return_value = client_mock
return boto_mock
@patch("boto3.DEFAULT_SESSION")
def test_default_session(boto3_default_session):
sess = Session()
assert sess.boto_session is boto3_default_session
@patch("boto3.Session")
def test_new_session_created(boto3_session):
sess = Session()
assert sess.boto_session is boto3_session.return_value
def test_process(boto_session):
session = Session(boto_session)
process_request_args = {
"inputs": [
{
"InputName": "input-1",
"S3Input": {
"S3Uri": "mocked_s3_uri_from_upload_data",
"LocalPath": "/container/path/",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
{
"InputName": "my_dataset",
"S3Input": {
"S3Uri": "s3://path/to/my/dataset/census.csv",
"LocalPath": "/container/path/",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
{
"InputName": "source",
"S3Input": {
"S3Uri": "mocked_s3_uri_from_upload_data",
"LocalPath": "/code/source",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
],
"output_config": {
"Outputs": [
{
"OutputName": "output-1",
"S3Output": {
"S3Uri": "s3://mybucket/current_job_name/output",
"LocalPath": "/data/output",
"S3UploadMode": "Continuous",
},
},
{
"OutputName": "my_output",
"S3Output": {
"S3Uri": "s3://uri/",
"LocalPath": "/container/path/",
"S3UploadMode": "Continuous",
},
},
],
"KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key",
},
"job_name": "current_job_name",
"resources": {
"ClusterConfig": {
"InstanceType": "ml.m4.xlarge",
"InstanceCount": 1,
"VolumeSizeInGB": 100,
}
},
"stopping_condition": {"MaxRuntimeInSeconds": 3600},
"app_specification": {
"ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3",
"ContainerArguments": ["--drop-columns", "'SelfEmployed'"],
"ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"],
},
"environment": {"my_env_variable": 20},
"network_config": {
"EnableInterContainerTrafficEncryption": True,
"EnableNetworkIsolation": True,
"VpcConfig": {
"SecurityGroupIds": ["my_security_group_id"],
"Subnets": ["my_subnet_id"],
},
},
"role_arn": ROLE,
"tags": [{"Name": "my-tag", "Value": "my-tag-value"}],
"experiment_config": {"ExperimentName": "AnExperiment"},
}
session.process(**process_request_args)
expected_request = {
"ProcessingJobName": "current_job_name",
"ProcessingResources": {
"ClusterConfig": {
"InstanceType": "ml.m4.xlarge",
"InstanceCount": 1,
"VolumeSizeInGB": 100,
}
},
"AppSpecification": {
"ImageUri": "520713654638.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:0.20.0-cpu-py3",
"ContainerArguments": ["--drop-columns", "'SelfEmployed'"],
"ContainerEntrypoint": ["python3", "/code/source/sklearn_transformer.py"],
},
"RoleArn": ROLE,
"ProcessingInputs": [
{
"InputName": "input-1",
"S3Input": {
"S3Uri": "mocked_s3_uri_from_upload_data",
"LocalPath": "/container/path/",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
{
"InputName": "my_dataset",
"S3Input": {
"S3Uri": "s3://path/to/my/dataset/census.csv",
"LocalPath": "/container/path/",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
{
"InputName": "source",
"S3Input": {
"S3Uri": "mocked_s3_uri_from_upload_data",
"LocalPath": "/code/source",
"S3DataType": "Archive",
"S3InputMode": "File",
"S3DataDistributionType": "FullyReplicated",
"S3CompressionType": "None",
},
},
],
"ProcessingOutputConfig": {
"Outputs": [
{
"OutputName": "output-1",
"S3Output": {
"S3Uri": "s3://mybucket/current_job_name/output",
"LocalPath": "/data/output",
"S3UploadMode": "Continuous",
},
},
{
"OutputName": "my_output",
"S3Output": {
"S3Uri": "s3://uri/",
"LocalPath": "/container/path/",
"S3UploadMode": "Continuous",
},
},
],
"KmsKeyId": "arn:aws:kms:us-west-2:012345678901:key/kms-key",
},
"Environment": {"my_env_variable": 20},
"NetworkConfig": {
"EnableInterContainerTrafficEncryption": True,
"EnableNetworkIsolation": True,
"VpcConfig": {
"SecurityGroupIds": ["my_security_group_id"],
"Subnets": ["my_subnet_id"],
},
},
"StoppingCondition": {"MaxRuntimeInSeconds": 3600},
"Tags": [{"Name": "my-tag", "Value": "my-tag-value"}],
"ExperimentConfig": {"ExperimentName": "AnExperiment"},
}
session.sagemaker_client.create_processing_job.assert_called_with(**expected_request)
def mock_exists(filepath_to_mock, exists_result):
unmocked_exists = os.path.exists
def side_effect(filepath):
if filepath == filepath_to_mock:
return exists_result
else:
return unmocked_exists(filepath)
return Mock(side_effect=side_effect)
def test_get_execution_role():
session = Mock()
session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:role/SageMakerRole"
actual = get_execution_role(session)
assert actual == "arn:aws:iam::369233609183:role/SageMakerRole"
def test_get_execution_role_works_with_service_role():
session = Mock()
session.get_caller_identity_arn.return_value = (
"arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388"
)
actual = get_execution_role(session)
assert (
actual
== "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388"
)
def test_get_execution_role_throws_exception_if_arn_is_not_role():
session = Mock()
session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos"
with pytest.raises(ValueError) as error:
get_execution_role(session)
assert "The current AWS identity is not a role" in str(error.value)
def test_get_execution_role_throws_exception_if_arn_is_not_role_with_role_in_name():
session = Mock()
session.get_caller_identity_arn.return_value = "arn:aws:iam::369233609183:user/marcos-role"
with pytest.raises(ValueError) as error:
get_execution_role(session)
assert "The current AWS identity is not a role" in str(error.value)
@patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}'))
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True))
def test_get_caller_identity_arn_from_describe_notebook_instance(boto_session):
sess = Session(boto_session)
expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388"
sess.sagemaker_client.describe_notebook_instance.return_value = {"RoleArn": expected_role}
actual = sess.get_caller_identity_arn()
assert actual == expected_role
sess.sagemaker_client.describe_notebook_instance.assert_called_once_with(
NotebookInstanceName="SageMakerInstance"
)
@patch(
"six.moves.builtins.open",
mock_open(
read_data='{"ResourceName": "SageMakerInstance", '
'"DomainId": "d-kbnw5yk6tg8j", '
'"UserProfileName": "default-1617915559064"}'
),
)
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True))
def test_get_caller_identity_arn_from_describe_user_profile(boto_session):
sess = Session(boto_session)
expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388"
sess.sagemaker_client.describe_user_profile.return_value = {
"UserSettings": {"ExecutionRole": expected_role}
}
actual = sess.get_caller_identity_arn()
assert actual == expected_role
sess.sagemaker_client.describe_user_profile.assert_called_once_with(
DomainId="d-kbnw5yk6tg8j",
UserProfileName="default-1617915559064",
)
@patch(
"six.moves.builtins.open",
mock_open(
read_data='{"ResourceName": "SageMakerInstance", '
'"DomainId": "d-kbnw5yk6tg8j", '
'"UserProfileName": "default-1617915559064"}'
),
)
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True))
def test_get_caller_identity_arn_from_describe_domain_if_no_user_settings(boto_session):
sess = Session(boto_session)
expected_role = "arn:aws:iam::369233609183:role/service-role/SageMakerRole-20171129T072388"
sess.sagemaker_client.describe_user_profile.return_value = {}
sess.sagemaker_client.describe_domain.return_value = {
"DefaultUserSettings": {"ExecutionRole": expected_role}
}
actual = sess.get_caller_identity_arn()
assert actual == expected_role
sess.sagemaker_client.describe_user_profile.assert_called_once_with(
DomainId="d-kbnw5yk6tg8j",
UserProfileName="default-1617915559064",
)
sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j")
@patch(
"six.moves.builtins.open",
mock_open(
read_data='{"ResourceName": "SageMakerInstance", '
'"DomainId": "d-kbnw5yk6tg8j", '
'"UserProfileName": "default-1617915559064"}'
),
)
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True))
def test_fallback_to_domain_if_role_unavailable_in_user_settings(boto_session):
sess = Session(boto_session)
expected_role = "expected_role"
sess.sagemaker_client.describe_user_profile.return_value = {
"DomainId": "d-kbnw5yk6tg8j",
"UserSettings": {
"JupyterServerAppSettings": {},
"KernelGatewayAppSettings": {},
},
}
sess.sagemaker_client.describe_domain.return_value = {
"DefaultUserSettings": {"ExecutionRole": expected_role}
}
actual = sess.get_caller_identity_arn()
assert actual == expected_role
sess.sagemaker_client.describe_user_profile.assert_called_once_with(
DomainId="d-kbnw5yk6tg8j",
UserProfileName="default-1617915559064",
)
sess.sagemaker_client.describe_domain.assert_called_once_with(DomainId="d-kbnw5yk6tg8j")
@patch("six.moves.builtins.open", mock_open(read_data='{"ResourceName": "SageMakerInstance"}'))
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, True))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_a_role_after_describe_notebook_exception(
sts_regional_endpoint, boto_session
):
sess = Session(boto_session)
exception = ClientError(
{"Error": {"Code": "ValidationException", "Message": "RecordNotFound"}}, "Operation"
)
sess.sagemaker_client.describe_notebook_instance.side_effect = exception
arn = (
"arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122"
)
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": arn
}
expected_role = "arn:aws:iam::369233609183:role/SageMakerRole"
sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}}
with patch("logging.Logger.debug") as mock_logger:
actual = sess.get_caller_identity_arn()
mock_logger.assert_called_once()
sess.sagemaker_client.describe_notebook_instance.assert_called_once_with(
NotebookInstanceName="SageMakerInstance"
)
assert actual == expected_role
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_a_user(sts_regional_endpoint, boto_session):
sess = Session(boto_session)
arn = "arn:aws:iam::369233609183:user/mia"
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": arn
}
sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": arn}}
actual = sess.get_caller_identity_arn()
assert actual == "arn:aws:iam::369233609183:user/mia"
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_an_user_without_permissions(
sts_regional_endpoint, boto_session
):
sess = Session(boto_session)
arn = "arn:aws:iam::369233609183:user/mia"
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": arn
}
sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {})
with patch("logging.Logger.warning") as mock_logger:
actual = sess.get_caller_identity_arn()
assert actual == "arn:aws:iam::369233609183:user/mia"
mock_logger.assert_called_once()
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_a_role(sts_regional_endpoint, boto_session):
sess = Session(boto_session)
arn = (
"arn:aws:sts::369233609183:assumed-role/SageMakerRole/6d009ef3-5306-49d5-8efc-78db644d8122"
)
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": arn
}
expected_role = "arn:aws:iam::369233609183:role/SageMakerRole"
sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": expected_role}}
actual = sess.get_caller_identity_arn()
assert actual == expected_role
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_an_execution_role(sts_regional_endpoint, boto_session):
sess = Session(boto_session)
sts_arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker"
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": sts_arn
}
iam_arn = "arn:aws:iam::369233609183:role/AmazonSageMaker-ExecutionRole-20171129T072388"
sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": iam_arn}}
actual = sess.get_caller_identity_arn()
assert actual == iam_arn
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_a_sagemaker_execution_role_with_iam_client_error(
sts_regional_endpoint, boto_session
):
sess = Session(boto_session)
arn = "arn:aws:sts::369233609183:assumed-role/AmazonSageMaker-ExecutionRole-20171129T072388/SageMaker"
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": arn
}
sess.boto_session.client("iam").get_role.side_effect = ClientError({}, {})
actual = sess.get_caller_identity_arn()
assert (
actual
== "arn:aws:iam::369233609183:role/service-role/AmazonSageMaker-ExecutionRole-20171129T072388"
)
@patch("os.path.exists", side_effect=mock_exists(NOTEBOOK_METADATA_FILE, False))
@patch("sagemaker.session.sts_regional_endpoint", return_value=STS_ENDPOINT)
def test_get_caller_identity_arn_from_role_with_path(sts_regional_endpoint, boto_session):
sess = Session(boto_session)
arn_prefix = "arn:aws:iam::369233609183:role"
role_name = "name"
sess.boto_session.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Arn": "/".join([arn_prefix, role_name])
}
role_path = "path"
role_with_path = "/".join([arn_prefix, role_path, role_name])
sess.boto_session.client("iam").get_role.return_value = {"Role": {"Arn": role_with_path}}
actual = sess.get_caller_identity_arn()
assert actual == role_with_path
def test_delete_endpoint(boto_session):
sess = Session(boto_session)
sess.delete_endpoint("my_endpoint")
boto_session.client().delete_endpoint.assert_called_with(EndpointName="my_endpoint")
def test_delete_endpoint_config(boto_session):
sess = Session(boto_session)
sess.delete_endpoint_config("my_endpoint_config")
boto_session.client().delete_endpoint_config.assert_called_with(
EndpointConfigName="my_endpoint_config"
)
def test_delete_model(boto_session):
sess = Session(boto_session)
model_name = "my_model"
sess.delete_model(model_name)
boto_session.client().delete_model.assert_called_with(ModelName=model_name)
def test_user_agent_injected(boto_session):
assert (
"AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent
)
sess = Session(boto_session)
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent
assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent
assert (
"AWS-SageMaker-Notebook-Instance"
not in sess.sagemaker_runtime_client._client_config.user_agent
)
def test_user_agent_injected_with_nbi(boto_session):
assert (
"AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent
)
with patch("six.moves.builtins.open", mock_open(read_data="120.0-0")) as mo:
sess = Session(boto_session)
mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt")
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent
assert "AWS-SageMaker-Notebook-Instance" in sess.sagemaker_client._client_config.user_agent
assert (
"AWS-SageMaker-Notebook-Instance" in sess.sagemaker_runtime_client._client_config.user_agent
)
def test_user_agent_injected_with_nbi_ioerror(boto_session):
assert (
"AWS-SageMaker-Python-SDK" not in boto_session.client("sagemaker")._client_config.user_agent
)
with patch("six.moves.builtins.open", MagicMock(side_effect=IOError("File not found"))) as mo:
sess = Session(boto_session)
mo.assert_called_with("/etc/opt/ml/sagemaker-notebook-instance-version.txt")
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_client._client_config.user_agent
assert "AWS-SageMaker-Python-SDK" in sess.sagemaker_runtime_client._client_config.user_agent
assert "AWS-SageMaker-Notebook-Instance" not in sess.sagemaker_client._client_config.user_agent
assert (
"AWS-SageMaker-Notebook-Instance"
not in sess.sagemaker_runtime_client._client_config.user_agent
)
def test_training_input_all_defaults():
prefix = "pre"
actual = TrainingInput(s3_data=prefix)
expected = {
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": prefix,
}
}
}
assert actual.config == expected
def test_training_input_all_arguments():
prefix = "pre"
distribution = "FullyReplicated"
compression = "Gzip"
content_type = "text/csv"
record_wrapping = "RecordIO"
s3_data_type = "Manifestfile"
input_mode = "Pipe"
result = TrainingInput(
s3_data=prefix,
distribution=distribution,
compression=compression,
input_mode=input_mode,
content_type=content_type,
record_wrapping=record_wrapping,
s3_data_type=s3_data_type,
)
expected = {
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": distribution,
"S3DataType": s3_data_type,
"S3Uri": prefix,
}
},
"CompressionType": compression,
"ContentType": content_type,
"RecordWrapperType": record_wrapping,
"InputMode": input_mode,
}
assert result.config == expected
IMAGE = "myimage"
S3_INPUT_URI = "s3://mybucket/data"
S3_OUTPUT = "s3://sagemaker-123/output/jobname"
ROLE = "SageMakerRole"
EXPANDED_ROLE = "arn:aws:iam::111111111111:role/ExpandedRole"
INSTANCE_COUNT = 1
INSTANCE_TYPE = "ml.c4.xlarge"
ACCELERATOR_TYPE = "ml.eia.medium"
MAX_SIZE = 30
MAX_TIME = 3 * 60 * 60
JOB_NAME = "jobname"
TAGS = [{"Name": "some-tag", "Value": "value-for-tag"}]
VPC_CONFIG = {"Subnets": ["foo"], "SecurityGroupIds": ["bar"]}
METRIC_DEFINITONS = [{"Name": "validation-rmse", "Regex": "validation-rmse=(\\d+)"}]
EXPERIMENT_CONFIG = {
"ExperimentName": "dummyExp",
"TrialName": "dummyT",
"TrialComponentDisplayName": "dummyTC",
}
MODEL_CLIENT_CONFIG = {"InvocationsMaxRetries": 2, "InvocationsTimeoutInSeconds": 60}
DEFAULT_EXPECTED_TRAIN_JOB_ARGS = {
"OutputDataConfig": {"S3OutputPath": S3_OUTPUT},
"RoleArn": EXPANDED_ROLE,
"ResourceConfig": {
"InstanceCount": INSTANCE_COUNT,
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": MAX_SIZE,
},
"InputDataConfig": [
{
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": S3_INPUT_URI,
}
},
"ChannelName": "training",
}
],
"AlgorithmSpecification": {"TrainingInputMode": "File", "TrainingImage": IMAGE},
"TrainingJobName": JOB_NAME,
"StoppingCondition": {"MaxRuntimeInSeconds": MAX_TIME},
"VpcConfig": VPC_CONFIG,
"ExperimentConfig": EXPERIMENT_CONFIG,
}
COMPLETED_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS)
COMPLETED_DESCRIBE_JOB_RESULT.update(
{"TrainingJobArn": "arn:aws:sagemaker:us-west-2:336:training-job/" + JOB_NAME}
)
COMPLETED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Completed"})
COMPLETED_DESCRIBE_JOB_RESULT.update(
{"ModelArtifacts": {"S3ModelArtifacts": S3_OUTPUT + "/model/model.tar.gz"}}
)
# TrainingStartTime and TrainingEndTime are for billable seconds calculation
COMPLETED_DESCRIBE_JOB_RESULT.update(
{"TrainingStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000)}
)
COMPLETED_DESCRIBE_JOB_RESULT.update(
{"TrainingEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000)}
)
STOPPED_DESCRIBE_JOB_RESULT = dict(COMPLETED_DESCRIBE_JOB_RESULT)
STOPPED_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "Stopped"})
IN_PROGRESS_DESCRIBE_JOB_RESULT = dict(DEFAULT_EXPECTED_TRAIN_JOB_ARGS)
IN_PROGRESS_DESCRIBE_JOB_RESULT.update({"TrainingJobStatus": "InProgress"})
COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT = {
"TransformJobStatus": "Completed",
"ModelName": "some-model",
"TransformJobName": JOB_NAME,
"TransformResources": {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE},
"TransformEndTime": datetime.datetime(2018, 2, 17, 7, 19, 34, 953000),
"TransformStartTime": datetime.datetime(2018, 2, 17, 7, 15, 0, 103000),
"TransformOutput": {"AssembleWith": "None", "KmsKeyId": "", "S3OutputPath": S3_OUTPUT},
"TransformInput": {
"CompressionType": "None",
"ContentType": "text/csv",
"DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI},
"SplitType": "Line",
},
}
STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT)
STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "Stopped"})
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT = dict(COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT)
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT.update({"TransformJobStatus": "InProgress"})
SERVERLESS_INFERENCE_CONFIG = {
"MemorySizeInMB": 2048,
"MaxConcurrency": 2,
}
@pytest.fixture()
def sagemaker_session():
boto_mock = MagicMock(name="boto_session")
boto_mock.client("sts", endpoint_url=STS_ENDPOINT).get_caller_identity.return_value = {
"Account": "123"
}
ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock())
ims.expand_role = Mock(return_value=EXPANDED_ROLE)
return ims
def test_train_pack_to_request(sagemaker_session):
in_config = [
{
"ChannelName": "training",
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": S3_INPUT_URI,
}
},
}
]
out_config = {"S3OutputPath": S3_OUTPUT}
resource_config = {
"InstanceCount": INSTANCE_COUNT,
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": MAX_SIZE,
}
stop_cond = {"MaxRuntimeInSeconds": MAX_TIME}
sagemaker_session.train(
image_uri=IMAGE,
input_mode="File",
input_config=in_config,
role=EXPANDED_ROLE,
job_name=JOB_NAME,
output_config=out_config,
resource_config=resource_config,
hyperparameters=None,
stop_condition=stop_cond,
tags=None,
vpc_config=VPC_CONFIG,
metric_definitions=None,
experiment_config=EXPERIMENT_CONFIG,
enable_sagemaker_metrics=None,
)
assert sagemaker_session.sagemaker_client.method_calls[0] == (
"create_training_job",
(),
DEFAULT_EXPECTED_TRAIN_JOB_ARGS,
)
SAMPLE_STOPPING_CONDITION = {"MaxRuntimeInSeconds": MAX_TIME}
RESOURCE_CONFIG = {
"InstanceCount": INSTANCE_COUNT,
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": MAX_SIZE,
}
SAMPLE_INPUT = [
{
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": S3_INPUT_URI,
}
},
"ChannelName": "training",
}
]
SAMPLE_OUTPUT = {"S3OutputPath": S3_OUTPUT}
SAMPLE_OBJECTIVE = {"Type": "Maximize", "MetricName": "val-score"}
SAMPLE_OBJECTIVE_2 = {"Type": "Maximize", "MetricName": "value-score"}
SAMPLE_METRIC_DEF = [{"Name": "train:progress", "Regex": "regex-1"}]
SAMPLE_METRIC_DEF_2 = [{"Name": "value-score", "Regex": "regex-2"}]
STATIC_HPs = {"feature_dim": "784"}
STATIC_HPs_2 = {"gamma": "0.1"}
SAMPLE_PARAM_RANGES = [{"Name": "mini_batch_size", "MinValue": "10", "MaxValue": "100"}]
SAMPLE_PARAM_RANGES_2 = [{"Name": "kernel", "Values": ["rbf", "sigmoid"]}]
SAMPLE_TUNING_JOB_REQUEST = {
"HyperParameterTuningJobName": "dummy-tuning-1",
"HyperParameterTuningJobConfig": {
"Strategy": "Bayesian",
"HyperParameterTuningJobObjective": SAMPLE_OBJECTIVE,
"ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5},
"ParameterRanges": SAMPLE_PARAM_RANGES,
"TrainingJobEarlyStoppingType": "Off",
},
"TrainingJobDefinition": {
"StaticHyperParameters": STATIC_HPs,
"AlgorithmSpecification": {
"TrainingImage": "dummy-image-1",
"TrainingInputMode": "File",
"MetricDefinitions": SAMPLE_METRIC_DEF,
},
"RoleArn": EXPANDED_ROLE,
"InputDataConfig": SAMPLE_INPUT,
"OutputDataConfig": SAMPLE_OUTPUT,
"ResourceConfig": RESOURCE_CONFIG,
"StoppingCondition": SAMPLE_STOPPING_CONDITION,
},
}
SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST = {
"HyperParameterTuningJobName": "dummy-tuning-1",
"HyperParameterTuningJobConfig": {
"Strategy": "Bayesian",
"ResourceLimits": {"MaxNumberOfTrainingJobs": 100, "MaxParallelTrainingJobs": 5},
"TrainingJobEarlyStoppingType": "Off",
},
"TrainingJobDefinitions": [
{
"DefinitionName": "estimator_1",
"TuningObjective": SAMPLE_OBJECTIVE,
"HyperParameterRanges": SAMPLE_PARAM_RANGES,
"StaticHyperParameters": STATIC_HPs,
"AlgorithmSpecification": {
"TrainingImage": "dummy-image-1",
"TrainingInputMode": "File",
"MetricDefinitions": SAMPLE_METRIC_DEF,
},
"RoleArn": EXPANDED_ROLE,
"InputDataConfig": SAMPLE_INPUT,
"OutputDataConfig": SAMPLE_OUTPUT,
"ResourceConfig": RESOURCE_CONFIG,
"StoppingCondition": SAMPLE_STOPPING_CONDITION,
},
{
"DefinitionName": "estimator_2",
"TuningObjective": SAMPLE_OBJECTIVE_2,
"HyperParameterRanges": SAMPLE_PARAM_RANGES_2,
"StaticHyperParameters": STATIC_HPs_2,
"AlgorithmSpecification": {
"TrainingImage": "dummy-image-2",
"TrainingInputMode": "File",
"MetricDefinitions": SAMPLE_METRIC_DEF_2,
},
"RoleArn": EXPANDED_ROLE,
"InputDataConfig": SAMPLE_INPUT,
"OutputDataConfig": SAMPLE_OUTPUT,
"ResourceConfig": RESOURCE_CONFIG,
"StoppingCondition": SAMPLE_STOPPING_CONDITION,
},
],
}
@pytest.mark.parametrize(
"warm_start_type, parents",
[("IdenticalDataAndAlgorithm", {"p1", "p2", "p3"}), ("TransferLearning", {"p1", "p2", "p3"})],
)
def test_tune_warm_start(sagemaker_session, warm_start_type, parents):
def assert_create_tuning_job_request(**kwrags):
assert (
kwrags["HyperParameterTuningJobConfig"]
== SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"]
)
assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"]
assert kwrags["WarmStartConfig"] == {
"WarmStartType": warm_start_type,
"ParentHyperParameterTuningJobs": [
{"HyperParameterTuningJobName": parent} for parent in parents
],
}
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.tune(
job_name="dummy-tuning-1",
strategy="Bayesian",
objective_type="Maximize",
objective_metric_name="val-score",
max_jobs=100,
max_parallel_jobs=5,
parameter_ranges=SAMPLE_PARAM_RANGES,
static_hyperparameters=STATIC_HPs,
image_uri="dummy-image-1",
input_mode="File",
metric_definitions=SAMPLE_METRIC_DEF,
role=EXPANDED_ROLE,
input_config=SAMPLE_INPUT,
output_config=SAMPLE_OUTPUT,
resource_config=RESOURCE_CONFIG,
stop_condition=SAMPLE_STOPPING_CONDITION,
tags=None,
warm_start_config=WarmStartConfig(
warm_start_type=WarmStartTypes(warm_start_type), parents=parents
).to_input_req(),
)
def test_create_tuning_job_without_training_config_or_list(sagemaker_session):
with pytest.raises(
ValueError, match="Either training_config or training_config_list should be provided."
):
sagemaker_session.create_tuning_job(
job_name="dummy-tuning-1",
tuning_config={
"strategy": "Bayesian",
"objective_type": "Maximize",
"objective_metric_name": "val-score",
"max_jobs": 100,
"max_parallel_jobs": 5,
"parameter_ranges": SAMPLE_PARAM_RANGES,
},
)
def test_create_tuning_job_with_both_training_config_and_list(sagemaker_session):
with pytest.raises(
ValueError, match="Only one of training_config and training_config_list should be provided."
):
sagemaker_session.create_tuning_job(
job_name="dummy-tuning-1",
tuning_config={
"strategy": "Bayesian",
"objective_type": "Maximize",
"objective_metric_name": "val-score",
"max_jobs": 100,
"max_parallel_jobs": 5,
"parameter_ranges": SAMPLE_PARAM_RANGES,
},
training_config={"static_hyperparameters": STATIC_HPs, "image_uri": "dummy-image-1"},
training_config_list=[
{
"static_hyperparameters": STATIC_HPs,
"image_uri": "dummy-image-1",
"estimator_name": "estimator_1",
},
{
"static_hyperparameters": STATIC_HPs_2,
"image_uri": "dummy-image-2",
"estimator_name": "estimator_2",
},
],
)
def test_create_tuning_job(sagemaker_session):
def assert_create_tuning_job_request(**kwrags):
assert (
kwrags["HyperParameterTuningJobConfig"]
== SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"]
)
assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"]
assert "TrainingJobDefinitions" not in kwrags
assert kwrags.get("WarmStartConfig", None) is None
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.create_tuning_job(
job_name="dummy-tuning-1",
tuning_config={
"strategy": "Bayesian",
"objective_type": "Maximize",
"objective_metric_name": "val-score",
"max_jobs": 100,
"max_parallel_jobs": 5,
"parameter_ranges": SAMPLE_PARAM_RANGES,
},
training_config={
"static_hyperparameters": STATIC_HPs,
"image_uri": "dummy-image-1",
"input_mode": "File",
"metric_definitions": SAMPLE_METRIC_DEF,
"role": EXPANDED_ROLE,
"input_config": SAMPLE_INPUT,
"output_config": SAMPLE_OUTPUT,
"resource_config": RESOURCE_CONFIG,
"stop_condition": SAMPLE_STOPPING_CONDITION,
},
tags=None,
warm_start_config=None,
)
def test_create_tuning_job_multi_algo(sagemaker_session):
def assert_create_tuning_job_request(**kwrags):
expected_tuning_config = SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST[
"HyperParameterTuningJobConfig"
]
assert kwrags["HyperParameterTuningJobConfig"] == expected_tuning_config
assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert "TrainingJobDefinition" not in kwrags
assert (
kwrags["TrainingJobDefinitions"]
== SAMPLE_MULTI_ALGO_TUNING_JOB_REQUEST["TrainingJobDefinitions"]
)
assert kwrags.get("WarmStartConfig", None) is None
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.create_tuning_job(
job_name="dummy-tuning-1",
tuning_config={"strategy": "Bayesian", "max_jobs": 100, "max_parallel_jobs": 5},
training_config_list=[
{
"static_hyperparameters": STATIC_HPs,
"image_uri": "dummy-image-1",
"input_mode": "File",
"metric_definitions": SAMPLE_METRIC_DEF,
"role": EXPANDED_ROLE,
"input_config": SAMPLE_INPUT,
"output_config": SAMPLE_OUTPUT,
"resource_config": RESOURCE_CONFIG,
"stop_condition": SAMPLE_STOPPING_CONDITION,
"estimator_name": "estimator_1",
"objective_type": "Maximize",
"objective_metric_name": "val-score",
"parameter_ranges": SAMPLE_PARAM_RANGES,
},
{
"static_hyperparameters": STATIC_HPs_2,
"image_uri": "dummy-image-2",
"input_mode": "File",
"metric_definitions": SAMPLE_METRIC_DEF_2,
"role": EXPANDED_ROLE,
"input_config": SAMPLE_INPUT,
"output_config": SAMPLE_OUTPUT,
"resource_config": RESOURCE_CONFIG,
"stop_condition": SAMPLE_STOPPING_CONDITION,
"estimator_name": "estimator_2",
"objective_type": "Maximize",
"objective_metric_name": "value-score",
"parameter_ranges": SAMPLE_PARAM_RANGES_2,
},
],
tags=None,
warm_start_config=None,
)
def test_tune(sagemaker_session):
def assert_create_tuning_job_request(**kwrags):
assert (
kwrags["HyperParameterTuningJobConfig"]
== SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"]
)
assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert kwrags["TrainingJobDefinition"] == SAMPLE_TUNING_JOB_REQUEST["TrainingJobDefinition"]
assert kwrags.get("WarmStartConfig", None) is None
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.tune(
job_name="dummy-tuning-1",
strategy="Bayesian",
objective_type="Maximize",
objective_metric_name="val-score",
max_jobs=100,
max_parallel_jobs=5,
parameter_ranges=SAMPLE_PARAM_RANGES,
static_hyperparameters=STATIC_HPs,
image_uri="dummy-image-1",
input_mode="File",
metric_definitions=SAMPLE_METRIC_DEF,
role=EXPANDED_ROLE,
input_config=SAMPLE_INPUT,
output_config=SAMPLE_OUTPUT,
resource_config=RESOURCE_CONFIG,
stop_condition=SAMPLE_STOPPING_CONDITION,
tags=None,
warm_start_config=None,
)
def test_tune_with_encryption_flag(sagemaker_session):
def assert_create_tuning_job_request(**kwrags):
assert (
kwrags["HyperParameterTuningJobConfig"]
== SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"]
)
assert kwrags["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert kwrags["TrainingJobDefinition"]["EnableInterContainerTrafficEncryption"] is True
assert kwrags.get("WarmStartConfig", None) is None
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.tune(
job_name="dummy-tuning-1",
strategy="Bayesian",
objective_type="Maximize",
objective_metric_name="val-score",
max_jobs=100,
max_parallel_jobs=5,
parameter_ranges=SAMPLE_PARAM_RANGES,
static_hyperparameters=STATIC_HPs,
image_uri="dummy-image-1",
input_mode="File",
metric_definitions=SAMPLE_METRIC_DEF,
role=EXPANDED_ROLE,
input_config=SAMPLE_INPUT,
output_config=SAMPLE_OUTPUT,
resource_config=RESOURCE_CONFIG,
stop_condition=SAMPLE_STOPPING_CONDITION,
tags=None,
warm_start_config=None,
encrypt_inter_container_traffic=True,
)
def test_tune_with_spot_and_checkpoints(sagemaker_session):
def assert_create_tuning_job_request(**kwargs):
assert (
kwargs["HyperParameterTuningJobConfig"]
== SAMPLE_TUNING_JOB_REQUEST["HyperParameterTuningJobConfig"]
)
assert kwargs["HyperParameterTuningJobName"] == "dummy-tuning-1"
assert kwargs["TrainingJobDefinition"]["EnableManagedSpotTraining"] is True
assert (
kwargs["TrainingJobDefinition"]["CheckpointConfig"]["S3Uri"]
== "s3://mybucket/checkpoints/"
)
assert (
kwargs["TrainingJobDefinition"]["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints"
)
assert kwargs.get("WarmStartConfig", None) is None
sagemaker_session.sagemaker_client.create_hyper_parameter_tuning_job.side_effect = (
assert_create_tuning_job_request
)
sagemaker_session.tune(
job_name="dummy-tuning-1",
strategy="Bayesian",
objective_type="Maximize",
objective_metric_name="val-score",
max_jobs=100,
max_parallel_jobs=5,
parameter_ranges=SAMPLE_PARAM_RANGES,
static_hyperparameters=STATIC_HPs,
image_uri="dummy-image-1",
input_mode="File",
metric_definitions=SAMPLE_METRIC_DEF,
role=EXPANDED_ROLE,
input_config=SAMPLE_INPUT,
output_config=SAMPLE_OUTPUT,
resource_config=RESOURCE_CONFIG,
stop_condition=SAMPLE_STOPPING_CONDITION,
tags=None,
warm_start_config=None,
use_spot_instances=True,
checkpoint_s3_uri="s3://mybucket/checkpoints/",
checkpoint_local_path="/tmp/checkpoints",
)
def test_stop_tuning_job(sagemaker_session):
sms = sagemaker_session
sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock(
name="stop_hyper_parameter_tuning_job"
)
sagemaker_session.stop_tuning_job(JOB_NAME)
sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with(
HyperParameterTuningJobName=JOB_NAME
)
def test_stop_tuning_job_client_error_already_stopped(sagemaker_session):
sms = sagemaker_session
exception = ClientError({"Error": {"Code": "ValidationException"}}, "Operation")
sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock(
name="stop_hyper_parameter_tuning_job", side_effect=exception
)
sagemaker_session.stop_tuning_job(JOB_NAME)
sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with(
HyperParameterTuningJobName=JOB_NAME
)
def test_stop_tuning_job_client_error(sagemaker_session):
error_response = {"Error": {"Code": "MockException", "Message": "MockMessage"}}
operation = "Operation"
exception = ClientError(error_response, operation)
sms = sagemaker_session
sms.sagemaker_client.stop_hyper_parameter_tuning_job = Mock(
name="stop_hyper_parameter_tuning_job", side_effect=exception
)
with pytest.raises(ClientError) as e:
sagemaker_session.stop_tuning_job(JOB_NAME)
sms.sagemaker_client.stop_hyper_parameter_tuning_job.assert_called_once_with(
HyperParameterTuningJobName=JOB_NAME
)
assert (
"An error occurred (MockException) when calling the Operation operation: MockMessage"
in str(e)
)
def test_train_pack_to_request_with_optional_params(sagemaker_session):
in_config = [
{
"ChannelName": "training",
"DataSource": {
"S3DataSource": {
"S3DataDistributionType": "FullyReplicated",
"S3DataType": "S3Prefix",
"S3Uri": S3_INPUT_URI,
}
},
}
]
out_config = {"S3OutputPath": S3_OUTPUT}
resource_config = {
"InstanceCount": INSTANCE_COUNT,
"InstanceType": INSTANCE_TYPE,
"VolumeSizeInGB": MAX_SIZE,
}
stop_cond = {"MaxRuntimeInSeconds": MAX_TIME}
RETRY_STRATEGY = {"MaximumRetryAttempts": 2}
hyperparameters = {"foo": "bar"}
sagemaker_session.train(
image_uri=IMAGE,
input_mode="File",
input_config=in_config,
role=EXPANDED_ROLE,
job_name=JOB_NAME,
output_config=out_config,
resource_config=resource_config,
vpc_config=VPC_CONFIG,
hyperparameters=hyperparameters,
stop_condition=stop_cond,
tags=TAGS,
metric_definitions=METRIC_DEFINITONS,
encrypt_inter_container_traffic=True,
use_spot_instances=True,
checkpoint_s3_uri="s3://mybucket/checkpoints/",
checkpoint_local_path="/tmp/checkpoints",
enable_sagemaker_metrics=True,
environment=ENV_INPUT,
retry_strategy=RETRY_STRATEGY,
)
_, _, actual_train_args = sagemaker_session.sagemaker_client.method_calls[0]
assert actual_train_args["VpcConfig"] == VPC_CONFIG
assert actual_train_args["HyperParameters"] == hyperparameters
assert actual_train_args["Tags"] == TAGS
assert actual_train_args["AlgorithmSpecification"]["MetricDefinitions"] == METRIC_DEFINITONS
assert actual_train_args["AlgorithmSpecification"]["EnableSageMakerMetricsTimeSeries"] is True
assert actual_train_args["EnableInterContainerTrafficEncryption"] is True
assert actual_train_args["EnableManagedSpotTraining"] is True
assert actual_train_args["CheckpointConfig"]["S3Uri"] == "s3://mybucket/checkpoints/"
assert actual_train_args["CheckpointConfig"]["LocalPath"] == "/tmp/checkpoints"
assert actual_train_args["Environment"] == ENV_INPUT
assert actual_train_args["RetryStrategy"] == RETRY_STRATEGY
def test_transform_pack_to_request(sagemaker_session):
model_name = "my-model"
in_config = {
"CompressionType": "None",
"ContentType": "text/csv",
"SplitType": "None",
"DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}},
}
out_config = {"S3OutputPath": S3_OUTPUT}
resource_config = {"InstanceCount": INSTANCE_COUNT, "InstanceType": INSTANCE_TYPE}
data_processing = {"OutputFilter": "$", "InputFilter": "$", "JoinSource": "Input"}
expected_args = {
"TransformJobName": JOB_NAME,
"ModelName": model_name,
"TransformInput": in_config,
"TransformOutput": out_config,
"TransformResources": resource_config,
"DataProcessing": data_processing,
}
sagemaker_session.transform(
job_name=JOB_NAME,
model_name=model_name,
strategy=None,
max_concurrent_transforms=None,
max_payload=None,
env=None,
input_config=in_config,
output_config=out_config,
resource_config=resource_config,
experiment_config=None,
model_client_config=None,
tags=None,
data_processing=data_processing,
batch_data_capture_config=None,
)
_, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0]
assert actual_args == expected_args
def test_transform_pack_to_request_with_optional_params(sagemaker_session):
strategy = "strategy"
max_concurrent_transforms = 1
max_payload = 0
env = {"FOO": "BAR"}
batch_data_capture_config = BatchDataCaptureConfig(
destination_s3_uri="test_uri",
kms_key_id="",
generate_inference_id=False,
)
sagemaker_session.transform(
job_name=JOB_NAME,
model_name="my-model",
strategy=strategy,
max_concurrent_transforms=max_concurrent_transforms,
env=env,
max_payload=max_payload,
input_config={},
output_config={},
resource_config={},
experiment_config=EXPERIMENT_CONFIG,
model_client_config=MODEL_CLIENT_CONFIG,
tags=TAGS,
data_processing=None,
batch_data_capture_config=batch_data_capture_config,
)
_, _, actual_args = sagemaker_session.sagemaker_client.method_calls[0]
assert actual_args["BatchStrategy"] == strategy
assert actual_args["MaxConcurrentTransforms"] == max_concurrent_transforms
assert actual_args["MaxPayloadInMB"] == max_payload
assert actual_args["Environment"] == env
assert actual_args["Tags"] == TAGS
assert actual_args["ExperimentConfig"] == EXPERIMENT_CONFIG
assert actual_args["ModelClientConfig"] == MODEL_CLIENT_CONFIG
assert actual_args["DataCaptureConfig"] == batch_data_capture_config._to_request_dict()
@patch("sys.stdout", new_callable=io.BytesIO if six.PY2 else io.StringIO)
def test_color_wrap(bio):
color_wrap = sagemaker.logs.ColorWrap()
color_wrap(0, "hi there")
assert bio.getvalue() == "hi there\n"
class MockBotoException(ClientError):
def __init__(self, code):
self.response = {"Error": {"Code": code}}
DEFAULT_LOG_STREAMS = {"logStreams": [{"logStreamName": JOB_NAME + "/xxxxxxxxx"}]}
LIFECYCLE_LOG_STREAMS = [
MockBotoException("ResourceNotFoundException"),
DEFAULT_LOG_STREAMS,
DEFAULT_LOG_STREAMS,
DEFAULT_LOG_STREAMS,
DEFAULT_LOG_STREAMS,
DEFAULT_LOG_STREAMS,
DEFAULT_LOG_STREAMS,
]
DEFAULT_LOG_EVENTS = [
{"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]},
{"nextForwardToken": None, "events": []},
]
STREAM_LOG_EVENTS = [
{"nextForwardToken": None, "events": [{"timestamp": 1, "message": "hi there #1"}]},
{"nextForwardToken": None, "events": []},
{
"nextForwardToken": None,
"events": [
{"timestamp": 1, "message": "hi there #1"},
{"timestamp": 2, "message": "hi there #2"},
],
},
{"nextForwardToken": None, "events": []},
{
"nextForwardToken": None,
"events": [
{"timestamp": 2, "message": "hi there #2"},
{"timestamp": 2, "message": "hi there #2a"},
{"timestamp": 3, "message": "hi there #3"},
],
},
{"nextForwardToken": None, "events": []},
]
@pytest.fixture()
def sagemaker_session_complete():
boto_mock = MagicMock(name="boto_session")
boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS
boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS
ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock())
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.sagemaker_client.describe_transform_job.return_value = (
COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT
)
return ims
@pytest.fixture()
def sagemaker_session_stopped():
boto_mock = MagicMock(name="boto_session")
boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS
boto_mock.client("logs").get_log_events.side_effect = DEFAULT_LOG_EVENTS
ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock())
ims.sagemaker_client.describe_training_job.return_value = STOPPED_DESCRIBE_JOB_RESULT
ims.sagemaker_client.describe_transform_job.return_value = STOPPED_DESCRIBE_TRANSFORM_JOB_RESULT
return ims
@pytest.fixture()
def sagemaker_session_ready_lifecycle():
boto_mock = MagicMock(name="boto_session")
boto_mock.client("logs").describe_log_streams.return_value = DEFAULT_LOG_STREAMS
boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS
ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock())
ims.sagemaker_client.describe_training_job.side_effect = [
IN_PROGRESS_DESCRIBE_JOB_RESULT,
IN_PROGRESS_DESCRIBE_JOB_RESULT,
COMPLETED_DESCRIBE_JOB_RESULT,
]
ims.sagemaker_client.describe_transform_job.side_effect = [
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT,
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT,
COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT,
]
return ims
@pytest.fixture()
def sagemaker_session_full_lifecycle():
boto_mock = MagicMock(name="boto_session")
boto_mock.client("logs").describe_log_streams.side_effect = LIFECYCLE_LOG_STREAMS
boto_mock.client("logs").get_log_events.side_effect = STREAM_LOG_EVENTS
ims = sagemaker.Session(boto_session=boto_mock, sagemaker_client=MagicMock())
ims.sagemaker_client.describe_training_job.side_effect = [
IN_PROGRESS_DESCRIBE_JOB_RESULT,
IN_PROGRESS_DESCRIBE_JOB_RESULT,
COMPLETED_DESCRIBE_JOB_RESULT,
]
ims.sagemaker_client.describe_transform_job.side_effect = [
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT,
IN_PROGRESS_DESCRIBE_TRANSFORM_JOB_RESULT,
COMPLETED_DESCRIBE_TRANSFORM_JOB_RESULT,
]
return ims
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_job_no_wait(cw, sagemaker_session_complete):
ims = sagemaker_session_complete
ims.logs_for_job(JOB_NAME)
ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME)
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_job_no_wait_stopped_job(cw, sagemaker_session_stopped):
ims = sagemaker_session_stopped
ims.logs_for_job(JOB_NAME)
ims.sagemaker_client.describe_training_job.assert_called_once_with(TrainingJobName=JOB_NAME)
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_job_wait_on_completed(cw, sagemaker_session_complete):
ims = sagemaker_session_complete
ims.logs_for_job(JOB_NAME, wait=True, poll=0)
assert ims.sagemaker_client.describe_training_job.call_args_list == [
call(TrainingJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_job_wait_on_stopped(cw, sagemaker_session_stopped):
ims = sagemaker_session_stopped
ims.logs_for_job(JOB_NAME, wait=True, poll=0)
assert ims.sagemaker_client.describe_training_job.call_args_list == [
call(TrainingJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle):
ims = sagemaker_session_ready_lifecycle
ims.logs_for_job(JOB_NAME)
assert ims.sagemaker_client.describe_training_job.call_args_list == [
call(TrainingJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
@patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180])
def test_logs_for_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle):
ims = sagemaker_session_full_lifecycle
ims.logs_for_job(JOB_NAME, wait=True, poll=0)
assert (
ims.sagemaker_client.describe_training_job.call_args_list
== [call(TrainingJobName=JOB_NAME)] * 3
)
assert cw().call_args_list == [
call(0, "hi there #1"),
call(0, "hi there #2"),
call(0, "hi there #2a"),
call(0, "hi there #3"),
]
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_transform_job_no_wait(cw, sagemaker_session_complete):
ims = sagemaker_session_complete
ims.logs_for_transform_job(JOB_NAME)
ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME)
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_transform_job_no_wait_stopped_job(cw, sagemaker_session_stopped):
ims = sagemaker_session_stopped
ims.logs_for_transform_job(JOB_NAME)
ims.sagemaker_client.describe_transform_job.assert_called_once_with(TransformJobName=JOB_NAME)
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_transform_job_wait_on_completed(cw, sagemaker_session_complete):
ims = sagemaker_session_complete
ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0)
assert ims.sagemaker_client.describe_transform_job.call_args_list == [
call(TransformJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_transform_job_wait_on_stopped(cw, sagemaker_session_stopped):
ims = sagemaker_session_stopped
ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0)
assert ims.sagemaker_client.describe_transform_job.call_args_list == [
call(TransformJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
def test_logs_for_transform_job_no_wait_on_running(cw, sagemaker_session_ready_lifecycle):
ims = sagemaker_session_ready_lifecycle
ims.logs_for_transform_job(JOB_NAME)
assert ims.sagemaker_client.describe_transform_job.call_args_list == [
call(TransformJobName=JOB_NAME)
]
cw().assert_called_with(0, "hi there #1")
@patch("sagemaker.logs.ColorWrap")
@patch("time.time", side_effect=[0, 30, 60, 90, 120, 150, 180])
def test_logs_for_transform_job_full_lifecycle(time, cw, sagemaker_session_full_lifecycle):
ims = sagemaker_session_full_lifecycle
ims.logs_for_transform_job(JOB_NAME, wait=True, poll=0)
assert (
ims.sagemaker_client.describe_transform_job.call_args_list
== [call(TransformJobName=JOB_NAME)] * 3
)
assert cw().call_args_list == [
call(0, "hi there #1"),
call(0, "hi there #2"),
call(0, "hi there #2a"),
call(0, "hi there #3"),
]
MODEL_NAME = "some-model"
PRIMARY_CONTAINER = {
"Environment": {},
"Image": IMAGE,
"ModelDataUrl": "s3://sagemaker-123/output/jobname/model/model.tar.gz",
}
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model(expand_container_def, sagemaker_session):
model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER)
assert model == MODEL_NAME
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_with_tags(expand_container_def, sagemaker_session):
tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}]
model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, tags=tags)
assert model == MODEL_NAME
tags = [{"Value": "TagtestValue", "Key": "TagtestKey"}]
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE,
ModelName=MODEL_NAME,
PrimaryContainer=PRIMARY_CONTAINER,
Tags=tags,
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_with_primary_container(expand_container_def, sagemaker_session):
model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER)
assert model == MODEL_NAME
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, PrimaryContainer=PRIMARY_CONTAINER
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_with_both(expand_container_def, sagemaker_session):
with pytest.raises(ValueError):
sagemaker_session.create_model(
MODEL_NAME, ROLE, container_defs=PRIMARY_CONTAINER, primary_container=PRIMARY_CONTAINER
)
CONTAINERS = [
{
"Environment": {"SAGEMAKER_DEFAULT_INVOCATIONS_ACCEPT": "application/json"},
"Image": "mi-1",
"ModelDataUrl": "s3://bucket/model_1.tar.gz",
"Framework": "TENSORFLOW",
"FrameworkVersion": "2.9",
"NearestModelName": "resnet50",
"ModelInput": {
"DataInputConfig": '{"input_1":[1,224,224,3]}',
},
},
{"Environment": {}, "Image": "mi-2", "ModelDataUrl": "s3://bucket/model_2.tar.gz"},
]
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_pipeline_model(expand_container_def, sagemaker_session):
model = sagemaker_session.create_model(MODEL_NAME, ROLE, container_defs=CONTAINERS)
assert model == MODEL_NAME
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE, ModelName=MODEL_NAME, Containers=CONTAINERS
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_vpc_config(expand_container_def, sagemaker_session):
model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER, VPC_CONFIG)
assert model == MODEL_NAME
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE,
ModelName=MODEL_NAME,
PrimaryContainer=PRIMARY_CONTAINER,
VpcConfig=VPC_CONFIG,
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_pipeline_model_vpc_config(expand_container_def, sagemaker_session):
model = sagemaker_session.create_model(MODEL_NAME, ROLE, CONTAINERS, VPC_CONFIG)
assert model == MODEL_NAME
sagemaker_session.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE,
ModelName=MODEL_NAME,
Containers=CONTAINERS,
VpcConfig=VPC_CONFIG,
)
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_already_exists(expand_container_def, sagemaker_session, caplog):
error_response = {
"Error": {"Code": "ValidationException", "Message": "Cannot create already existing model"}
}
exception = ClientError(error_response, "Operation")
sagemaker_session.sagemaker_client.create_model.side_effect = exception
model = sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER)
assert model == MODEL_NAME
expected_warning = (
"sagemaker",
logging.WARNING,
"Using already existing model: {}".format(MODEL_NAME),
)
assert expected_warning in caplog.record_tuples
@patch("sagemaker.session._expand_container_def", return_value=PRIMARY_CONTAINER)
def test_create_model_failure(expand_container_def, sagemaker_session):
error_message = "this is expected"
sagemaker_session.sagemaker_client.create_model.side_effect = RuntimeError(error_message)
with pytest.raises(RuntimeError) as e:
sagemaker_session.create_model(MODEL_NAME, ROLE, PRIMARY_CONTAINER)
assert error_message in str(e)
def test_create_model_from_job(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.create_model_from_job(JOB_NAME)
assert (
call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list
)
ims.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE,
ModelName=JOB_NAME,
PrimaryContainer=PRIMARY_CONTAINER,
VpcConfig=VPC_CONFIG,
)
def test_create_model_from_job_with_tags(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.create_model_from_job(JOB_NAME, tags=TAGS)
assert (
call(TrainingJobName=JOB_NAME) in ims.sagemaker_client.describe_training_job.call_args_list
)
ims.sagemaker_client.create_model.assert_called_with(
ExecutionRoleArn=EXPANDED_ROLE,
ModelName=JOB_NAME,
PrimaryContainer=PRIMARY_CONTAINER,
VpcConfig=VPC_CONFIG,
Tags=TAGS,
)
def test_create_model_from_job_with_image(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.create_model_from_job(JOB_NAME, image_uri="some-image")
[create_model_call] = ims.sagemaker_client.create_model.call_args_list
assert dict(create_model_call[1]["PrimaryContainer"])["Image"] == "some-image"
def test_create_model_from_job_with_container_def(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.create_model_from_job(
JOB_NAME,
image_uri="some-image",
model_data_url="some-data",
env={"a": "b"},
)
[create_model_call] = ims.sagemaker_client.create_model.call_args_list
c_def = create_model_call[1]["PrimaryContainer"]
assert c_def["Image"] == "some-image"
assert c_def["ModelDataUrl"] == "some-data"
assert c_def["Environment"] == {"a": "b"}
def test_create_model_from_job_with_vpc_config_override(sagemaker_session):
vpc_config_override = {"Subnets": ["foo", "bar"], "SecurityGroupIds": ["baz"]}
ims = sagemaker_session
ims.sagemaker_client.describe_training_job.return_value = COMPLETED_DESCRIBE_JOB_RESULT
ims.create_model_from_job(JOB_NAME, vpc_config_override=vpc_config_override)
assert ims.sagemaker_client.create_model.call_args[1]["VpcConfig"] == vpc_config_override
ims.create_model_from_job(JOB_NAME, vpc_config_override=None)
assert "VpcConfig" not in ims.sagemaker_client.create_model.call_args[1]
def test_endpoint_from_production_variants(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"})
pvs = [
sagemaker.production_variant("A", "ml.p2.xlarge"),
sagemaker.production_variant("B", "p299.4096xlarge"),
]
ex = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b"
)
ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex)
sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs)
sagemaker_session.sagemaker_client.create_endpoint.assert_called_with(
EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[]
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="some-endpoint", ProductionVariants=pvs
)
def test_create_endpoint_config_with_tags(sagemaker_session):
tags = [{"Key": "TagtestKey", "Value": "TagtestValue"}]
sagemaker_session.create_endpoint_config("endpoint-test", "simple-model", 1, "local", tags=tags)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="endpoint-test", ProductionVariants=ANY, Tags=tags
)
def test_endpoint_from_production_variants_with_tags(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"})
pvs = [
sagemaker.production_variant("A", "ml.p2.xlarge"),
sagemaker.production_variant("B", "p299.4096xlarge"),
]
ex = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b"
)
ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex)
tags = [{"ModelName": "TestModel"}]
sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags)
sagemaker_session.sagemaker_client.create_endpoint.assert_called_with(
EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags
)
def test_endpoint_from_production_variants_with_accelerator_type(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"})
pvs = [
sagemaker.production_variant("A", "ml.p2.xlarge", accelerator_type=ACCELERATOR_TYPE),
sagemaker.production_variant("B", "p299.4096xlarge", accelerator_type=ACCELERATOR_TYPE),
]
ex = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b"
)
ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex)
tags = [{"ModelName": "TestModel"}]
sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags)
sagemaker_session.sagemaker_client.create_endpoint.assert_called_with(
EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags
)
def test_endpoint_from_production_variants_with_serverless_inference_config(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"})
pvs = [
sagemaker.production_variant(
"A", "ml.p2.xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG
),
sagemaker.production_variant(
"B", "p299.4096xlarge", serverless_inference_config=SERVERLESS_INFERENCE_CONFIG
),
]
ex = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b"
)
ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex)
tags = [{"ModelName": "TestModel"}]
sagemaker_session.endpoint_from_production_variants("some-endpoint", pvs, tags)
sagemaker_session.sagemaker_client.create_endpoint.assert_called_with(
EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=tags
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="some-endpoint", ProductionVariants=pvs, Tags=tags
)
def test_endpoint_from_production_variants_with_async_config(sagemaker_session):
ims = sagemaker_session
ims.sagemaker_client.describe_endpoint = Mock(return_value={"EndpointStatus": "InService"})
pvs = [
sagemaker.production_variant("A", "ml.p2.xlarge"),
sagemaker.production_variant("B", "p299.4096xlarge"),
]
ex = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find your thing"}}, "b"
)
ims.sagemaker_client.describe_endpoint_config = Mock(side_effect=ex)
sagemaker_session.endpoint_from_production_variants(
"some-endpoint",
pvs,
async_inference_config_dict=AsyncInferenceConfig,
)
sagemaker_session.sagemaker_client.create_endpoint.assert_called_with(
EndpointConfigName="some-endpoint", EndpointName="some-endpoint", Tags=[]
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName="some-endpoint",
ProductionVariants=pvs,
AsyncInferenceConfig=AsyncInferenceConfig,
)
def test_update_endpoint_succeed(sagemaker_session):
sagemaker_session.sagemaker_client.describe_endpoint = Mock(
return_value={"EndpointStatus": "InService"}
)
endpoint_name = "some-endpoint"
endpoint_config = "some-endpoint-config"
returned_endpoint_name = sagemaker_session.update_endpoint(endpoint_name, endpoint_config)
assert returned_endpoint_name == endpoint_name
def test_update_endpoint_no_wait(sagemaker_session):
sagemaker_session.sagemaker_client.describe_endpoint = Mock(
return_value={"EndpointStatus": "Updating"}
)
endpoint_name = "some-endpoint"
endpoint_config = "some-endpoint-config"
returned_endpoint_name = sagemaker_session.update_endpoint(
endpoint_name, endpoint_config, wait=False
)
assert returned_endpoint_name == endpoint_name
def test_update_endpoint_non_existing_endpoint(sagemaker_session):
error = ClientError(
{"Error": {"Code": "ValidationException", "Message": "Could not find entity"}}, "foo"
)
expected_error_message = (
"Endpoint with name 'non-existing-endpoint' does not exist; "
"please use an existing endpoint name"
)
sagemaker_session.sagemaker_client.describe_endpoint = Mock(side_effect=error)
with pytest.raises(ValueError, match=expected_error_message):
sagemaker_session.update_endpoint("non-existing-endpoint", "non-existing-config")
def test_create_endpoint_config_from_existing(sagemaker_session):
pvs = [sagemaker.production_variant("A", "ml.m4.xlarge")]
tags = [{"Key": "aws:cloudformation:stackname", "Value": "this-tag-should-be-ignored"}]
existing_endpoint_arn = "arn:aws:sagemaker:us-west-2:123412341234:endpoint-config/foo"
kms_key = "kms"
sagemaker_session.sagemaker_client.describe_endpoint_config.return_value = {
"Tags": tags,
"ProductionVariants": pvs,
"EndpointConfigArn": existing_endpoint_arn,
"KmsKeyId": kms_key,
}
sagemaker_session.sagemaker_client.list_tags.return_value = {"Tags": tags}
existing_endpoint_name = "foo"
new_endpoint_name = "new-foo"
sagemaker_session.create_endpoint_config_from_existing(
existing_endpoint_name, new_endpoint_name
)
sagemaker_session.sagemaker_client.describe_endpoint_config.assert_called_with(
EndpointConfigName=existing_endpoint_name
)
sagemaker_session.sagemaker_client.list_tags.assert_called_with(
ResourceArn=existing_endpoint_arn, MaxResults=50
)
sagemaker_session.sagemaker_client.create_endpoint_config.assert_called_with(
EndpointConfigName=new_endpoint_name, ProductionVariants=pvs, KmsKeyId=kms_key
)
@patch("time.sleep")
def test_wait_for_tuning_job(sleep, sagemaker_session):
hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock(
name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc
)
result = sagemaker_session.wait_for_tuning_job(JOB_NAME)
assert result["HyperParameterTuningJobStatus"] == "Completed"
def test_tune_job_status(sagemaker_session):
hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock(
name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc
)
result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME)
assert result["HyperParameterTuningJobStatus"] == "Completed"
def test_tune_job_status_none(sagemaker_session):
hyperparameter_tuning_job_desc = {"HyperParameterTuningJobStatus": "InProgress"}
sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job = Mock(
name="describe_hyper_parameter_tuning_job", return_value=hyperparameter_tuning_job_desc
)
result = _tuning_job_status(sagemaker_session.sagemaker_client, JOB_NAME)
assert result is None
@patch("time.sleep")
def test_wait_for_transform_job_completed(sleep, sagemaker_session):
transform_job_desc = {"TransformJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_transform_job = Mock(
name="describe_transform_job", return_value=transform_job_desc
)
assert sagemaker_session.wait_for_transform_job(JOB_NAME)["TransformJobStatus"] == "Completed"
@patch("time.sleep")
def test_wait_for_transform_job_in_progress(sleep, sagemaker_session):
transform_job_desc_in_progress = {"TransformJobStatus": "InProgress"}
transform_job_desc_in_completed = {"TransformJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_transform_job = Mock(
name="describe_transform_job",
side_effect=[transform_job_desc_in_progress, transform_job_desc_in_completed],
)
assert (
sagemaker_session.wait_for_transform_job(JOB_NAME, 1)["TransformJobStatus"] == "Completed"
)
assert 2 == sagemaker_session.sagemaker_client.describe_transform_job.call_count
def test_transform_job_status(sagemaker_session):
transform_job_desc = {"TransformJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_transform_job = Mock(
name="describe_transform_job", return_value=transform_job_desc
)
result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME)
assert result["TransformJobStatus"] == "Completed"
def test_transform_job_status_none(sagemaker_session):
transform_job_desc = {"TransformJobStatus": "InProgress"}
sagemaker_session.sagemaker_client.describe_transform_job = Mock(
name="describe_transform_job", return_value=transform_job_desc
)
result = _transform_job_status(sagemaker_session.sagemaker_client, JOB_NAME)
assert result is None
def test_train_done_completed(sagemaker_session):
training_job_desc = {"TrainingJobStatus": "Completed"}
sagemaker_session.sagemaker_client.describe_training_job = Mock(
name="describe_training_job", return_value=training_job_desc
)
actual_job_desc, training_finished = _train_done(
sagemaker_session.sagemaker_client, JOB_NAME, None
)
assert actual_job_desc["TrainingJobStatus"] == "Completed"
assert training_finished is True
def test_train_done_in_progress(sagemaker_session):
training_job_desc = {"TrainingJobStatus": "InProgress"}
sagemaker_session.sagemaker_client.describe_training_job = Mock(
name="describe_training_job", return_value=training_job_desc
)
actual_job_desc, training_finished = _train_done(
sagemaker_session.sagemaker_client, JOB_NAME, None
)
assert actual_job_desc["TrainingJobStatus"] == "InProgress"
assert training_finished is False
DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS = {
"AutoMLJobName": JOB_NAME,
"InputDataConfig": [
{
"DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}},
"TargetAttributeName": "y",
}
],
"OutputDataConfig": {"S3OutputPath": S3_OUTPUT},
"AutoMLJobConfig": {
"CompletionCriteria": {
"MaxCandidates": 10,
"MaxAutoMLJobRuntimeInSeconds": 36000,
"MaxRuntimePerTrainingJobInSeconds": 3600 * 2,
}
},
"RoleArn": EXPANDED_ROLE,
"GenerateCandidateDefinitionsOnly": False,
}
COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS = {
"AutoMLJobName": JOB_NAME,
"InputDataConfig": [
{
"DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}},
"CompressionType": "Gzip",
"TargetAttributeName": "y",
}
],
"OutputDataConfig": {"S3OutputPath": S3_OUTPUT},
"ProblemType": "Regression",
"AutoMLJobObjective": {"Type": "type", "MetricName": "metric-name"},
"AutoMLJobConfig": {
"CompletionCriteria": {
"MaxCandidates": 10,
"MaxAutoMLJobRuntimeInSeconds": 36000,
"MaxRuntimePerTrainingJobInSeconds": 3600 * 2,
},
"SecurityConfig": {
"VolumeKmsKeyId": "volume-kms-key-id-string",
"EnableInterContainerTrafficEncryption": False,
"VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]},
},
},
"RoleArn": EXPANDED_ROLE,
"GenerateCandidateDefinitionsOnly": True,
"Tags": ["tag"],
}
COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS = {
"AutoMLJobName": JOB_NAME,
"StatusEquals": "Completed",
"SortOrder": "Descending",
"SortBy": "Status",
"MaxResults": 10,
}
def test_auto_ml_pack_to_request(sagemaker_session):
input_config = [
{
"DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}},
"TargetAttributeName": "y",
}
]
output_config = {"S3OutputPath": S3_OUTPUT}
auto_ml_job_config = {
"CompletionCriteria": {
"MaxCandidates": 10,
"MaxAutoMLJobRuntimeInSeconds": 36000,
"MaxRuntimePerTrainingJobInSeconds": 3600 * 2,
}
}
job_name = JOB_NAME
role = EXPANDED_ROLE
sagemaker_session.auto_ml(input_config, output_config, auto_ml_job_config, role, job_name)
assert sagemaker_session.sagemaker_client.method_calls[0] == (
"create_auto_ml_job",
(),
DEFAULT_EXPECTED_AUTO_ML_JOB_ARGS,
)
def test_auto_ml_pack_to_request_with_optional_args(sagemaker_session):
input_config = [
{
"DataSource": {"S3DataSource": {"S3DataType": "S3Prefix", "S3Uri": S3_INPUT_URI}},
"CompressionType": "Gzip",
"TargetAttributeName": "y",
}
]
output_config = {"S3OutputPath": S3_OUTPUT}
auto_ml_job_config = {
"CompletionCriteria": {
"MaxCandidates": 10,
"MaxAutoMLJobRuntimeInSeconds": 36000,
"MaxRuntimePerTrainingJobInSeconds": 3600 * 2,
},
"SecurityConfig": {
"VolumeKmsKeyId": "volume-kms-key-id-string",
"EnableInterContainerTrafficEncryption": False,
"VpcConfig": {"SecurityGroupIds": ["security-group-id"], "Subnets": ["subnet"]},
},
}
job_name = JOB_NAME
role = EXPANDED_ROLE
sagemaker_session.auto_ml(
input_config,
output_config,
auto_ml_job_config,
role,
job_name,
problem_type="Regression",
job_objective={"Type": "type", "MetricName": "metric-name"},
generate_candidate_definitions_only=True,
tags=["tag"],
)
assert sagemaker_session.sagemaker_client.method_calls[0] == (
"create_auto_ml_job",
(),
COMPLETE_EXPECTED_AUTO_ML_JOB_ARGS,
)
def test_list_candidates_for_auto_ml_job_default(sagemaker_session):
sagemaker_session.list_candidates(job_name=JOB_NAME)
sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once()
sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with(
AutoMLJobName=JOB_NAME
)
def test_list_candidates_for_auto_ml_job_with_optional_args(sagemaker_session):
sagemaker_session.list_candidates(
job_name=JOB_NAME,
status_equals="Completed",
sort_order="Descending",
sort_by="Status",
max_results=10,
)
sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_once()
sagemaker_session.sagemaker_client.list_candidates_for_auto_ml_job.assert_called_with(
**COMPLETE_EXPECTED_LIST_CANDIDATES_ARGS
)
def test_describe_tuning_job(sagemaker_session):
job_name = "hyper-parameter-tuning"
sagemaker_session.describe_tuning_job(job_name=job_name)
sagemaker_session.sagemaker_client.describe_hyper_parameter_tuning_job.assert_called_with(
HyperParameterTuningJobName=job_name
)
@pytest.fixture
def pipeline_empty_definition():
return '{"Version": "2020-12-01", ' '"Metadata": {}, ' '"Parameters": [], ' '"Steps": []}'
@pytest.fixture
def pipeline_role_arn():
return "my:pipeline:role:arn"
def test_describe_model(sagemaker_session):
model_name = "sagemaker-model-name"
sagemaker_session.describe_model(name=model_name)
sagemaker_session.sagemaker_client.describe_model.assert_called_with(ModelName=model_name)
def test_create_model_package_from_containers(sagemaker_session):
model_package_name = "sagemaker-model-package"
sagemaker_session.create_model_package_from_containers(model_package_name=model_package_name)
sagemaker_session.sagemaker_client.create_model_package.assert_called_once()
def test_create_model_package_from_containers_name_conflict(sagemaker_session):
model_package_name = "sagemaker-model-package"
model_package_group_name = "sagemaker-model-package-group"
with pytest.raises(ValueError) as error:
sagemaker_session.create_model_package_from_containers(
model_package_name=model_package_name,
model_package_group_name=model_package_group_name,
)
assert (
"model_package_name and model_package_group_name cannot be present at the same "
"time." == str(error)
)
def test_create_model_package_from_containers_incomplete_args(sagemaker_session):
model_package_name = "sagemaker-model-package"
containers = ["dummy-container"]
with pytest.raises(ValueError) as error:
sagemaker_session.create_model_package_from_containers(
model_package_name=model_package_name,
containers=containers,
)
assert (
"content_types and response_types "
"must be provided if containers is present." == str(error)
)
def test_create_model_package_from_containers_without_model_package_group_name(sagemaker_session):
model_package_name = "sagemaker-model-package"
containers = ["dummy-container"]
content_types = ["application/json"]
response_types = ["application/json"]
with pytest.raises(ValueError) as error:
sagemaker_session.create_model_package_from_containers(
model_package_name=model_package_name,
containers=containers,
content_types=content_types,
response_types=response_types,
)
assert (
"inference_inferences and transform_instances "
"must be provided if model_package_group_name is not present." == str(error)
)
def test_create_model_package_from_containers_all_args(sagemaker_session):
model_package_name = "sagemaker-model-package"
containers = ["dummy-container"]
content_types = ["application/json"]
response_types = ["application/json"]
inference_instances = ["ml.m4.xlarge"]
transform_instances = ["ml.m4.xlarget"]
model_metrics = {
"Bias": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
drift_check_baselines = {
"Bias": {
"ConfigFile": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
}
metadata_properties = {
"CommitId": "test-commit-id",
"Repository": "test-repository",
"GeneratedBy": "sagemaker-python-sdk",
"ProjectId": "unit-test",
}
marketplace_cert = (True,)
approval_status = ("Approved",)
description = "description"
customer_metadata_properties = {"key1": "value1"}
domain = "COMPUTER_VISION"
task = "IMAGE_CLASSIFICATION"
sample_payload_url = "s3://test-bucket/model"
sagemaker_session.create_model_package_from_containers(
containers=containers,
content_types=content_types,
response_types=response_types,
inference_instances=inference_instances,
transform_instances=transform_instances,
model_package_name=model_package_name,
model_metrics=model_metrics,
metadata_properties=metadata_properties,
marketplace_cert=marketplace_cert,
approval_status=approval_status,
description=description,
drift_check_baselines=drift_check_baselines,
customer_metadata_properties=customer_metadata_properties,
domain=domain,
sample_payload_url=sample_payload_url,
task=task,
)
expected_args = {
"ModelPackageName": model_package_name,
"InferenceSpecification": {
"Containers": containers,
"SupportedContentTypes": content_types,
"SupportedResponseMIMETypes": response_types,
"SupportedRealtimeInferenceInstanceTypes": inference_instances,
"SupportedTransformInstanceTypes": transform_instances,
},
"ModelPackageDescription": description,
"ModelMetrics": model_metrics,
"MetadataProperties": metadata_properties,
"CertifyForMarketplace": marketplace_cert,
"ModelApprovalStatus": approval_status,
"DriftCheckBaselines": drift_check_baselines,
"CustomerMetadataProperties": customer_metadata_properties,
"Domain": domain,
"SamplePayloadUrl": sample_payload_url,
"Task": task,
}
sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args)
def test_create_model_package_from_containers_without_instance_types(sagemaker_session):
model_package_group_name = "sagemaker-model-package-group-name-1.0"
containers = ["dummy-container"]
content_types = ["application/json"]
response_types = ["application/json"]
model_metrics = {
"Bias": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
drift_check_baselines = {
"Bias": {
"ConfigFile": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
}
metadata_properties = {
"CommitId": "test-commit-id",
"Repository": "test-repository",
"GeneratedBy": "sagemaker-python-sdk",
"ProjectId": "unit-test",
}
marketplace_cert = (True,)
approval_status = ("Approved",)
description = "description"
customer_metadata_properties = {"key1": "value1"}
sagemaker_session.create_model_package_from_containers(
containers=containers,
content_types=content_types,
response_types=response_types,
model_package_group_name=model_package_group_name,
model_metrics=model_metrics,
metadata_properties=metadata_properties,
marketplace_cert=marketplace_cert,
approval_status=approval_status,
description=description,
drift_check_baselines=drift_check_baselines,
customer_metadata_properties=customer_metadata_properties,
)
expected_args = {
"ModelPackageGroupName": model_package_group_name,
"InferenceSpecification": {
"Containers": containers,
"SupportedContentTypes": content_types,
"SupportedResponseMIMETypes": response_types,
},
"ModelPackageDescription": description,
"ModelMetrics": model_metrics,
"MetadataProperties": metadata_properties,
"CertifyForMarketplace": marketplace_cert,
"ModelApprovalStatus": approval_status,
"DriftCheckBaselines": drift_check_baselines,
"CustomerMetadataProperties": customer_metadata_properties,
}
sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args)
def test_create_model_package_from_containers_with_one_instance_types(sagemaker_session):
model_package_group_name = "sagemaker-model-package-group-name-1.0"
containers = ["dummy-container"]
content_types = ["application/json"]
response_types = ["application/json"]
transform_instances = ["ml.m5.xlarge"]
model_metrics = {
"Bias": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
drift_check_baselines = {
"Bias": {
"ConfigFile": {
"ContentType": "content-type",
"S3Uri": "s3://...",
}
}
}
metadata_properties = {
"CommitId": "test-commit-id",
"Repository": "test-repository",
"GeneratedBy": "sagemaker-python-sdk",
"ProjectId": "unit-test",
}
marketplace_cert = (True,)
approval_status = ("Approved",)
description = "description"
customer_metadata_properties = {"key1": "value1"}
sagemaker_session.create_model_package_from_containers(
containers=containers,
content_types=content_types,
response_types=response_types,
transform_instances=transform_instances,
model_package_group_name=model_package_group_name,
model_metrics=model_metrics,
metadata_properties=metadata_properties,
marketplace_cert=marketplace_cert,
approval_status=approval_status,
description=description,
drift_check_baselines=drift_check_baselines,
customer_metadata_properties=customer_metadata_properties,
)
expected_args = {
"ModelPackageGroupName": model_package_group_name,
"InferenceSpecification": {
"Containers": containers,
"SupportedContentTypes": content_types,
"SupportedResponseMIMETypes": response_types,
"SupportedTransformInstanceTypes": transform_instances,
},
"ModelPackageDescription": description,
"ModelMetrics": model_metrics,
"MetadataProperties": metadata_properties,
"CertifyForMarketplace": marketplace_cert,
"ModelApprovalStatus": approval_status,
"DriftCheckBaselines": drift_check_baselines,
"CustomerMetadataProperties": customer_metadata_properties,
}
sagemaker_session.sagemaker_client.create_model_package.assert_called_with(**expected_args)
@pytest.fixture
def feature_group_dummy_definitions():
return [{"FeatureName": "feature1", "FeatureType": "String"}]
def test_feature_group_create(sagemaker_session, feature_group_dummy_definitions):
sagemaker_session.create_feature_group(
feature_group_name="MyFeatureGroup",
record_identifier_name="feature1",
event_time_feature_name="feature2",
feature_definitions=feature_group_dummy_definitions,
role_arn="dummy_role",
)
assert sagemaker_session.sagemaker_client.create_feature_group.called_with(
FeatureGroupName="MyFeatureGroup",
RecordIdentifierFeatureName="feature1",
EventTimeFeatureName="feature2",
FeatureDefinitions=feature_group_dummy_definitions,
RoleArn="dummy_role",
)
def test_feature_group_delete(sagemaker_session):
sagemaker_session.delete_feature_group(feature_group_name="MyFeatureGroup")
assert sagemaker_session.sagemaker_client.delete_feature_group.called_with(
FeatureGroupName="MyFeatureGroup",
)
def test_feature_group_describe(sagemaker_session):
sagemaker_session.describe_feature_group(feature_group_name="MyFeatureGroup")
assert sagemaker_session.sagemaker_client.describe_feature_group.called_with(
FeatureGroupName="MyFeatureGroup",
)
def test_feature_group_update(sagemaker_session, feature_group_dummy_definitions):
sagemaker_session.update_feature_group(
feature_group_name="MyFeatureGroup",
feature_additions=feature_group_dummy_definitions,
)
assert sagemaker_session.sagemaker_client.update_feature_group.called_with(
FeatureGroupName="MyFeatureGroup",
FeatureAdditions=feature_group_dummy_definitions,
)
def test_feature_metadata_update(sagemaker_session):
parameter_additions = [
{
"key": "TestKey",
"value": "TestValue",
}
]
parameter_removals = ["TestKey"]
sagemaker_session.update_feature_metadata(
feature_group_name="TestFeatureGroup",
feature_name="TestFeature",
description="TestDescription",
parameter_additions=parameter_additions,
parameter_removals=parameter_removals,
)
assert sagemaker_session.sagemaker_client.update_feature_group.called_with(
feature_group_name="TestFeatureGroup",
FeatureName="TestFeature",
Description="TestDescription",
ParameterAdditions=parameter_additions,
ParameterRemovals=parameter_removals,
)
sagemaker_session.update_feature_metadata(
feature_group_name="TestFeatureGroup",
feature_name="TestFeature",
)
assert sagemaker_session.sagemaker_client.update_feature_group.called_with(
feature_group_name="TestFeatureGroup",
FeatureName="TestFeature",
)
def test_feature_metadata_describe(sagemaker_session):
sagemaker_session.describe_feature_metadata(
feature_group_name="MyFeatureGroup", feature_name="TestFeature"
)
assert sagemaker_session.sagemaker_client.describe_feature_metadata.called_with(
FeatureGroupName="MyFeatureGroup", FeatureName="TestFeature"
)
def test_start_query_execution(sagemaker_session):
athena_mock = Mock()
sagemaker_session.boto_session.client(
"athena", region_name=sagemaker_session.boto_region_name
).return_value = athena_mock
sagemaker_session.start_query_execution(
catalog="catalog",
database="database",
query_string="query",
output_location="s3://results",
)
assert athena_mock.start_query_execution.called_once_with(
QueryString="query",
QueryExecutionContext={"Catalog": "catalog", "Database": "database"},
OutputLocation="s3://results",
)
def test_get_query_execution(sagemaker_session):
athena_mock = Mock()
sagemaker_session.boto_session.client(
"athena", region_name=sagemaker_session.boto_region_name
).return_value = athena_mock
sagemaker_session.get_query_execution(query_execution_id="query_id")
assert athena_mock.get_query_execution.called_with(QueryExecutionId="query_id")
def test_download_athena_query_result(sagemaker_session):
sagemaker_session.s3_client = Mock()
sagemaker_session.download_athena_query_result(
bucket="bucket",
prefix="prefix",
query_execution_id="query_id",
filename="filename",
)
assert sagemaker_session.s3_client.download_file.called_with(
Bucket="bucket",
Key="prefix/query_id.csv",
Filename="filename",
)
@patch("sagemaker.session.Session.get_query_execution")
def test_wait_for_athena_query(query_execution, sagemaker_session):
query_execution.return_value = {"QueryExecution": {"Status": {"State": "SUCCEEDED"}}}
sagemaker_session.wait_for_athena_query(query_execution_id="query_id")
assert query_execution.called_with(query_execution_id="query_id")