| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from __future__ import absolute_import |
| |
|
| | import json |
| | import os |
| |
|
| | import pytest |
| | from tests.integ import DATA_DIR, TRANSFORM_DEFAULT_TIMEOUT_MINUTES |
| | from tests.integ.timeout import ( |
| | timeout_and_delete_endpoint_by_name, |
| | timeout_and_delete_model_with_transformer, |
| | ) |
| |
|
| | from sagemaker import image_uris |
| | from sagemaker.model import Model |
| | from sagemaker.pipeline import PipelineModel |
| | from sagemaker.predictor import Predictor |
| | from sagemaker.serializers import JSONSerializer |
| | from sagemaker.sparkml.model import SparkMLModel |
| | from sagemaker.utils import sagemaker_timestamp |
| |
|
| | SPARKML_DATA_PATH = os.path.join(DATA_DIR, "sparkml_model") |
| | XGBOOST_DATA_PATH = os.path.join(DATA_DIR, "xgboost_model") |
| | SPARKML_XGBOOST_DATA_DIR = "sparkml_xgboost_pipeline" |
| | VALID_DATA_PATH = os.path.join(DATA_DIR, SPARKML_XGBOOST_DATA_DIR, "valid_input.csv") |
| | INVALID_DATA_PATH = os.path.join(DATA_DIR, SPARKML_XGBOOST_DATA_DIR, "invalid_input.csv") |
| | SCHEMA = json.dumps( |
| | { |
| | "input": [ |
| | {"name": "Pclass", "type": "float"}, |
| | {"name": "Embarked", "type": "string"}, |
| | {"name": "Age", "type": "float"}, |
| | {"name": "Fare", "type": "float"}, |
| | {"name": "SibSp", "type": "float"}, |
| | {"name": "Sex", "type": "string"}, |
| | ], |
| | "output": {"name": "features", "struct": "vector", "type": "double"}, |
| | } |
| | ) |
| |
|
| |
|
| | def test_inference_pipeline_batch_transform(sagemaker_session, cpu_instance_type): |
| | sparkml_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(SPARKML_DATA_PATH, "mleap_model.tar.gz"), |
| | key_prefix="integ-test-data/sparkml/model", |
| | ) |
| | xgb_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(XGBOOST_DATA_PATH, "xgb_model.tar.gz"), |
| | key_prefix="integ-test-data/xgboost/model", |
| | ) |
| | batch_job_name = "test-inference-pipeline-batch-{}".format(sagemaker_timestamp()) |
| | sparkml_model = SparkMLModel( |
| | model_data=sparkml_model_data, |
| | env={"SAGEMAKER_SPARKML_SCHEMA": SCHEMA}, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| | xgb_image = image_uris.retrieve( |
| | "xgboost", sagemaker_session.boto_region_name, version="1", image_scope="inference" |
| | ) |
| | xgb_model = Model( |
| | model_data=xgb_model_data, image_uri=xgb_image, sagemaker_session=sagemaker_session |
| | ) |
| | model = PipelineModel( |
| | models=[sparkml_model, xgb_model], |
| | role="SageMakerRole", |
| | sagemaker_session=sagemaker_session, |
| | name=batch_job_name, |
| | ) |
| | transformer = model.transformer(1, cpu_instance_type) |
| | transform_input_key_prefix = "integ-test-data/sparkml_xgboost/transform" |
| | transform_input = transformer.sagemaker_session.upload_data( |
| | path=VALID_DATA_PATH, key_prefix=transform_input_key_prefix |
| | ) |
| |
|
| | with timeout_and_delete_model_with_transformer( |
| | transformer, sagemaker_session, minutes=TRANSFORM_DEFAULT_TIMEOUT_MINUTES |
| | ): |
| | transformer.transform(transform_input, content_type="text/csv", job_name=batch_job_name) |
| | transformer.wait() |
| |
|
| |
|
| | @pytest.mark.release |
| | @pytest.mark.skip( |
| | reason="This test has always failed, but the failure was masked by a bug. " |
| | "This test should be fixed. Details in https://github.com/aws/sagemaker-python-sdk/pull/968" |
| | ) |
| | def test_inference_pipeline_model_deploy(sagemaker_session, cpu_instance_type): |
| | sparkml_data_path = os.path.join(DATA_DIR, "sparkml_model") |
| | xgboost_data_path = os.path.join(DATA_DIR, "xgboost_model") |
| | endpoint_name = "test-inference-pipeline-deploy-{}".format(sagemaker_timestamp()) |
| | sparkml_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(sparkml_data_path, "mleap_model.tar.gz"), |
| | key_prefix="integ-test-data/sparkml/model", |
| | ) |
| | xgb_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(xgboost_data_path, "xgb_model.tar.gz"), |
| | key_prefix="integ-test-data/xgboost/model", |
| | ) |
| |
|
| | with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| | sparkml_model = SparkMLModel( |
| | model_data=sparkml_model_data, |
| | env={"SAGEMAKER_SPARKML_SCHEMA": SCHEMA}, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| | xgb_image = image_uris.retrieve( |
| | "xgboost", sagemaker_session.boto_region_name, version="1", image_scope="inference" |
| | ) |
| | xgb_model = Model( |
| | model_data=xgb_model_data, image_uri=xgb_image, sagemaker_session=sagemaker_session |
| | ) |
| | model = PipelineModel( |
| | models=[sparkml_model, xgb_model], |
| | role="SageMakerRole", |
| | sagemaker_session=sagemaker_session, |
| | name=endpoint_name, |
| | ) |
| | model.deploy(1, cpu_instance_type, endpoint_name=endpoint_name) |
| | predictor = Predictor( |
| | endpoint_name=endpoint_name, |
| | sagemaker_session=sagemaker_session, |
| | serializer=JSONSerializer, |
| | content_type="text/csv", |
| | accept="text/csv", |
| | ) |
| |
|
| | with open(VALID_DATA_PATH, "r") as f: |
| | valid_data = f.read() |
| | assert predictor.predict(valid_data) == "0.714013934135" |
| |
|
| | with open(INVALID_DATA_PATH, "r") as f: |
| | invalid_data = f.read() |
| | assert predictor.predict(invalid_data) is None |
| |
|
| | model.delete_model() |
| | with pytest.raises(Exception) as exception: |
| | sagemaker_session.sagemaker_client.describe_model(ModelName=model.name) |
| | assert "Could not find model" in str(exception.value) |
| |
|
| |
|
| | @pytest.mark.slow_test |
| | def test_inference_pipeline_model_deploy_and_update_endpoint( |
| | sagemaker_session, cpu_instance_type, alternative_cpu_instance_type |
| | ): |
| | sparkml_data_path = os.path.join(DATA_DIR, "sparkml_model") |
| | xgboost_data_path = os.path.join(DATA_DIR, "xgboost_model") |
| | endpoint_name = "test-inference-pipeline-deploy-{}".format(sagemaker_timestamp()) |
| | sparkml_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(sparkml_data_path, "mleap_model.tar.gz"), |
| | key_prefix="integ-test-data/sparkml/model", |
| | ) |
| | xgb_model_data = sagemaker_session.upload_data( |
| | path=os.path.join(xgboost_data_path, "xgb_model.tar.gz"), |
| | key_prefix="integ-test-data/xgboost/model", |
| | ) |
| |
|
| | with timeout_and_delete_endpoint_by_name(endpoint_name, sagemaker_session): |
| | sparkml_model = SparkMLModel( |
| | model_data=sparkml_model_data, |
| | env={"SAGEMAKER_SPARKML_SCHEMA": SCHEMA}, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| | xgb_image = image_uris.retrieve( |
| | "xgboost", sagemaker_session.boto_region_name, version="1", image_scope="inference" |
| | ) |
| | xgb_model = Model( |
| | model_data=xgb_model_data, image_uri=xgb_image, sagemaker_session=sagemaker_session |
| | ) |
| | model = PipelineModel( |
| | models=[sparkml_model, xgb_model], |
| | role="SageMakerRole", |
| | predictor_cls=Predictor, |
| | sagemaker_session=sagemaker_session, |
| | ) |
| | predictor = model.deploy(1, alternative_cpu_instance_type, endpoint_name=endpoint_name) |
| | endpoint_desc = sagemaker_session.sagemaker_client.describe_endpoint( |
| | EndpointName=endpoint_name |
| | ) |
| | old_config_name = endpoint_desc["EndpointConfigName"] |
| |
|
| | predictor.update_endpoint(initial_instance_count=1, instance_type=cpu_instance_type) |
| |
|
| | endpoint_desc = sagemaker_session.sagemaker_client.describe_endpoint( |
| | EndpointName=endpoint_name |
| | ) |
| | new_config_name = endpoint_desc["EndpointConfigName"] |
| | new_config = sagemaker_session.sagemaker_client.describe_endpoint_config( |
| | EndpointConfigName=new_config_name |
| | ) |
| |
|
| | assert old_config_name != new_config_name |
| | assert new_config["ProductionVariants"][0]["InstanceType"] == cpu_instance_type |
| | assert new_config["ProductionVariants"][0]["InitialInstanceCount"] == 1 |
| |
|
| | model.delete_model() |
| | with pytest.raises(Exception) as exception: |
| | sagemaker_session.sagemaker_client.describe_model(ModelName=model.name) |
| | assert "Could not find model" in str(exception.value) |
| |
|