| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Placeholder docstring""" |
| from __future__ import absolute_import, print_function |
|
|
| import json |
| import logging |
| import os |
| import re |
| import uuid |
| from abc import ABCMeta, abstractmethod |
| from typing import Any, Dict, Union, Optional, List |
|
|
| from six import string_types, with_metaclass |
| from six.moves.urllib.parse import urlparse |
|
|
| import sagemaker |
| from sagemaker import git_utils, image_uris, vpc_utils |
| from sagemaker.analytics import TrainingJobAnalytics |
| from sagemaker.debugger import ( |
| DEBUGGER_FLAG, |
| DebuggerHookConfig, |
| FrameworkProfile, |
| ProfilerConfig, |
| ProfilerRule, |
| Rule, |
| TensorBoardOutputConfig, |
| get_default_profiler_rule, |
| get_rule_container_image_uri, |
| RuleBase, |
| ) |
| from sagemaker.deprecations import removed_function, removed_kwargs, renamed_kwargs |
| from sagemaker.fw_utils import ( |
| UploadedCode, |
| _region_supports_debugger, |
| _region_supports_profiler, |
| get_mp_parameters, |
| tar_and_upload_dir, |
| validate_source_dir, |
| validate_source_code_input_against_pipeline_variables, |
| ) |
| from sagemaker.inputs import TrainingInput, FileSystemInput |
| from sagemaker.instance_group import InstanceGroup |
| from sagemaker.job import _Job |
| from sagemaker.jumpstart.utils import ( |
| add_jumpstart_tags, |
| get_jumpstart_base_name_if_jumpstart_model, |
| update_inference_tags_with_jumpstart_training_tags, |
| ) |
| from sagemaker.local import LocalSession |
| from sagemaker.model import ( |
| CONTAINER_LOG_LEVEL_PARAM_NAME, |
| DIR_PARAM_NAME, |
| JOB_NAME_PARAM_NAME, |
| NEO_ALLOWED_FRAMEWORKS, |
| SAGEMAKER_REGION_PARAM_NAME, |
| SCRIPT_PARAM_NAME, |
| Model, |
| ) |
| from sagemaker.predictor import Predictor |
| from sagemaker.s3 import S3Uploader, parse_s3_url |
| from sagemaker.session import Session |
| from sagemaker.transformer import Transformer |
| from sagemaker.utils import ( |
| base_from_name, |
| base_name_from_image, |
| build_dict, |
| get_config_value, |
| name_from_base, |
| to_string, |
| ) |
| from sagemaker.workflow import is_pipeline_variable |
| from sagemaker.workflow.entities import PipelineVariable |
| from sagemaker.workflow.pipeline_context import ( |
| PipelineSession, |
| runnable_by_pipeline, |
| ) |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| class EstimatorBase(with_metaclass(ABCMeta, object)): |
| """Handle end-to-end Amazon SageMaker training and deployment tasks. |
| |
| For introduction to model training and deployment, see |
| http://docs.aws.amazon.com/sagemaker/latest/dg/how-it-works-training.html |
| |
| Subclasses must define a way to determine what image to use for training, |
| what hyperparameters to use, and how to create an appropriate predictor |
| instance. |
| """ |
|
|
| LAUNCH_PT_XLA_ENV_NAME = "sagemaker_pytorch_xla_multi_worker_enabled" |
| LAUNCH_PS_ENV_NAME = "sagemaker_parameter_server_enabled" |
| LAUNCH_MPI_ENV_NAME = "sagemaker_mpi_enabled" |
| LAUNCH_SM_DDP_ENV_NAME = "sagemaker_distributed_dataparallel_enabled" |
| INSTANCE_TYPE = "sagemaker_instance_type" |
| MPI_NUM_PROCESSES_PER_HOST = "sagemaker_mpi_num_of_processes_per_host" |
| MPI_CUSTOM_MPI_OPTIONS = "sagemaker_mpi_custom_mpi_options" |
| SM_DDP_CUSTOM_MPI_OPTIONS = "sagemaker_distributed_dataparallel_custom_mpi_options" |
| CONTAINER_CODE_CHANNEL_SOURCEDIR_PATH = "/opt/ml/input/data/code/sourcedir.tar.gz" |
| JOB_CLASS_NAME = "training-job" |
|
|
| def __init__( |
| self, |
| role: str, |
| instance_count: Optional[Union[int, PipelineVariable]] = None, |
| instance_type: Optional[Union[str, PipelineVariable]] = None, |
| keep_alive_period_in_seconds: Optional[Union[int, PipelineVariable]] = None, |
| volume_size: Union[int, PipelineVariable] = 30, |
| volume_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| max_run: Union[int, PipelineVariable] = 24 * 60 * 60, |
| input_mode: Union[str, PipelineVariable] = "File", |
| output_path: Optional[Union[str, PipelineVariable]] = None, |
| output_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| base_job_name: Optional[str] = None, |
| sagemaker_session: Optional[Session] = None, |
| tags: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| subnets: Optional[List[Union[str, PipelineVariable]]] = None, |
| security_group_ids: Optional[List[Union[str, PipelineVariable]]] = None, |
| model_uri: Optional[str] = None, |
| model_channel_name: Union[str, PipelineVariable] = "model", |
| metric_definitions: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| encrypt_inter_container_traffic: Union[bool, PipelineVariable] = False, |
| use_spot_instances: Union[bool, PipelineVariable] = False, |
| max_wait: Optional[Union[int, PipelineVariable]] = None, |
| checkpoint_s3_uri: Optional[Union[str, PipelineVariable]] = None, |
| checkpoint_local_path: Optional[Union[str, PipelineVariable]] = None, |
| rules: Optional[List[RuleBase]] = None, |
| debugger_hook_config: Optional[Union[bool, DebuggerHookConfig]] = None, |
| tensorboard_output_config: Optional[TensorBoardOutputConfig] = None, |
| enable_sagemaker_metrics: Optional[Union[bool, PipelineVariable]] = None, |
| enable_network_isolation: Union[bool, PipelineVariable] = False, |
| profiler_config: Optional[ProfilerConfig] = None, |
| disable_profiler: bool = False, |
| environment: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| max_retry_attempts: Optional[Union[int, PipelineVariable]] = None, |
| source_dir: Optional[Union[str, PipelineVariable]] = None, |
| git_config: Optional[Dict[str, str]] = None, |
| hyperparameters: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| container_log_level: Union[int, PipelineVariable] = logging.INFO, |
| code_location: Optional[str] = None, |
| entry_point: Optional[Union[str, PipelineVariable]] = None, |
| dependencies: Optional[List[Union[str]]] = None, |
| instance_groups: Optional[List[InstanceGroup]] = None, |
| **kwargs, |
| ): |
| """Initialize an ``EstimatorBase`` instance. |
| |
| Args: |
| role (str): An AWS IAM role (either name or full ARN). The Amazon |
| SageMaker training jobs and APIs that create Amazon SageMaker |
| endpoints use this role to access training data and model |
| artifacts. After the endpoint is created, the inference code |
| might use the IAM role, if it needs to access an AWS resource. |
| instance_count (int or PipelineVariable): Number of Amazon EC2 instances to use |
| for training. Required if instance_groups is not set. |
| instance_type (str or PipelineVariable): Type of EC2 instance to use for training, |
| for example, ``'ml.c4.xlarge'``. Required if instance_groups is |
| not set. |
| keep_alive_period_in_seconds (int): The duration of time in seconds |
| to retain configured resources in a warm pool for subsequent |
| training jobs (default: None). |
| volume_size (int or PipelineVariable): Size in GB of the storage volume to use for |
| storing input and output data during training (default: 30). |
| |
| Must be large enough to store training data if File mode is |
| used, which is the default mode. |
| |
| When you use an ML instance with the EBS-only storage option |
| such as ``ml.c5`` and ``ml.p2``, |
| you must define the size of the EBS |
| volume through the ``volume_size`` parameter in the estimator class. |
| |
| .. note:: |
| |
| When you use an ML instance with `NVMe SSD volumes |
| <https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ssd-instance-store.html#nvme-ssd-volumes>`_ |
| such as ``ml.p4d``, ``ml.g4dn``, and ``ml.g5``, |
| do not include this parameter in the estimator configuration. |
| If you use one of those ML instance types, |
| SageMaker doesn't provision Amazon EBS General Purpose SSD |
| (gp2) storage nor take this parameter to adjust the NVMe instance storage. |
| Available storage is fixed to the NVMe instance storage |
| capacity. SageMaker configures storage paths for training |
| datasets, checkpoints, model artifacts, and outputs to use the |
| entire capacity of the instance storage. |
| |
| Note that if you include this parameter and specify a number that |
| exceeds the size of the NVMe volume attached to the instance type, |
| SageMaker returns an ``Invalid VolumeSizeInGB`` error. |
| |
| To look up instance types and their instance storage types |
| and volumes, see `Amazon EC2 Instance Types |
| <http://aws.amazon.com/ec2/instance-types/>`_. |
| |
| To find the default local paths defined by the SageMaker |
| training platform, see `Amazon SageMaker Training Storage |
| Folders for Training Datasets, Checkpoints, Model Artifacts, |
| and Outputs |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/model-train-storage.html>`_. |
| volume_kms_key (str or PipelineVariable): Optional. KMS key ID for encrypting EBS |
| volume attached to the training instance (default: None). |
| max_run (int or PipelineVariable): Timeout in seconds for training (default: 24 * |
| 60 * 60). After this amount of time Amazon SageMaker terminates |
| the job regardless of its current status. |
| input_mode (str or PipelineVariable): The input mode that the algorithm supports |
| (default: 'File'). Valid modes: |
| 'File' - Amazon SageMaker copies the training dataset from the |
| S3 location to a local directory. |
| 'Pipe' - Amazon SageMaker streams data directly from S3 to the |
| container via a Unix-named pipe. |
| 'FastFile' - Amazon SageMaker streams data from S3 on demand instead of |
| downloading the entire dataset before training begins. This argument can |
| be overriden on a per-channel basis using |
| ``sagemaker.inputs.TrainingInput.input_mode``. |
| output_path (str or PipelineVariable): S3 location for saving the training result (model |
| artifacts and output files). If not specified, results are |
| stored to a default bucket. If the bucket with the specific name |
| does not exist, the estimator creates the bucket during the |
| :meth:`~sagemaker.estimator.EstimatorBase.fit` method execution. |
| file:// urls are used for local mode. For example: 'file://model/' |
| will save to the model folder in the current directory. |
| output_kms_key (str or PipelineVariable): Optional. KMS key ID for encrypting the |
| training output (default: Your IAM role's KMS key for Amazon S3). |
| If you don't provide a KMS key ID, Amazon SageMaker uses the |
| default KMS key for Amazon S3 of the account linked to your |
| IAM role. |
| base_job_name (str): Prefix for training job name when the |
| :meth:`~sagemaker.estimator.EstimatorBase.fit` method launches. |
| If not specified, the estimator generates a default job name |
| based on the training image name and current timestamp. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the estimator creates one |
| using the default AWS configuration chain. |
| tags (list[dict[str, str] or list[dict[str, PipelineVariable]]): |
| List of tags for labeling a training job. For more, see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_Tag.html. |
| subnets (list[str] or list[PipelineVariable]): List of subnet ids. If not |
| specified training job will be created without VPC config. |
| security_group_ids (list[str] or list[PipelineVariable]): List of security group ids. |
| If not specified training job will be created without VPC config. |
| model_uri (str): URI where a pre-trained model is stored, either |
| locally or in S3 (default: None). If specified, the estimator |
| will create a channel pointing to the model so the training job |
| can download it. This model can be a 'model.tar.gz' from a |
| previous training job, or other artifacts coming from a |
| different source. |
| |
| In local mode, this should point to the path in which the model |
| is located and not the file itself, as local Docker containers |
| will try to mount the URI as a volume. |
| |
| More information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/cdf-training.html#td-deserialization |
| model_channel_name (str or PipelineVariable): Name of the channel where 'model_uri' will |
| be downloaded (default: 'model'). |
| metric_definitions (list[dict[str, str] or list[dict[str, PipelineVariable]]): |
| A list of dictionaries that defines the metric(s) used to evaluate the |
| training jobs. Each dictionary contains two keys: 'Name' for the name of the metric, |
| and 'Regex' for the regular expression used to extract the |
| metric from the logs. This should be defined only for jobs that |
| don't use an Amazon algorithm. |
| encrypt_inter_container_traffic (bool or PipelineVariable): Specifies whether traffic |
| between training containers is encrypted for the training job |
| (default: ``False``). |
| use_spot_instances (bool or PipelineVariable): Specifies whether to use SageMaker |
| Managed Spot instances for training. If enabled then the |
| ``max_wait`` arg should also be set. |
| More information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/model-managed-spot-training.html |
| (default: ``False``). |
| max_wait (int or PipelineVariable): Timeout in seconds waiting for spot training |
| job (default: None). After this amount of time Amazon |
| SageMaker will stop waiting for managed spot training job to |
| complete (default: None). |
| checkpoint_s3_uri (str or PipelineVariable): The S3 URI in which to persist checkpoints |
| that the algorithm persists (if any) during training. (default: |
| ``None``). |
| checkpoint_local_path (str or PipelineVariable): The local path that the algorithm |
| writes its checkpoints to. SageMaker will persist all files |
| under this path to `checkpoint_s3_uri` continually during |
| training. On job startup the reverse happens - data from the |
| s3 location is downloaded to this path before the algorithm is |
| started. If the path is unset then SageMaker assumes the |
| checkpoints will be provided under `/opt/ml/checkpoints/`. |
| (default: None). |
| rules (list[:class:`~sagemaker.debugger.RuleBase`]): A list of |
| :class:`~sagemaker.debugger.RuleBase` objects used to define |
| SageMaker Debugger rules for real-time analysis |
| (default: ``None``). For more information, |
| see `Continuous analyses through rules |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html |
| #continuous-analyses-through-rules)>`_. |
| debugger_hook_config (:class:`~sagemaker.debugger.DebuggerHookConfig` or bool): |
| Configuration for how debugging information is emitted with |
| SageMaker Debugger. If not specified, a default one is created using |
| the estimator's ``output_path``, unless the region does not |
| support SageMaker Debugger. To disable SageMaker Debugger, |
| set this parameter to ``False``. For more information, see |
| `Capture real-time debugging data during model training in Amazon SageMaker |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html# |
| capture-real-time-debugging-data-during-model-training-in-amazon-sagemaker>`_. |
| tensorboard_output_config (:class:`~sagemaker.debugger.TensorBoardOutputConfig`): |
| Configuration for customizing debugging visualization using TensorBoard |
| (default: None). For more information, |
| see `Capture real time tensorboard data |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html# |
| capture-real-time-tensorboard-data-from-the-debugging-hook>`_. |
| enable_sagemaker_metrics (bool or PipelineVariable): enable SageMaker Metrics Time |
| Series. For more information, see `AlgorithmSpecification API |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/ |
| API_AlgorithmSpecification.html#SageMaker-Type-AlgorithmSpecification- |
| EnableSageMakerMetricsTimeSeries>`_. |
| (default: None). |
| enable_network_isolation (bool or PipelineVariable): Specifies whether container will |
| run in network isolation mode (default: ``False``). Network |
| isolation mode restricts the container access to outside networks |
| (such as the Internet). The container does not make any inbound or |
| outbound network calls. Also known as Internet-free mode. |
| profiler_config (:class:`~sagemaker.debugger.ProfilerConfig`): |
| Configuration for how SageMaker Debugger collects |
| monitoring and profiling information from your training job. |
| If not specified, a default configuration is created using |
| the estimator's ``output_path``, unless the region does not |
| support SageMaker Debugger. To disable SageMaker Debugger |
| monitoring and profiling, set the |
| ``disable_profiler`` parameter to ``True``. |
| disable_profiler (bool): Specifies whether Debugger monitoring and profiling |
| will be disabled (default: ``False``). |
| environment (dict[str, str] or dict[str, PipelineVariable]) : Environment variables |
| to be set for use during training job (default: None) |
| max_retry_attempts (int or PipelineVariable): The number of times to move a job |
| to the STARTING status. You can specify between 1 and 30 attempts. |
| If the value of attempts is greater than zero, |
| the job is retried on InternalServerFailure |
| the same number of attempts as the value. |
| You can cap the total duration for your job by setting ``max_wait`` and ``max_run`` |
| (default: None) |
| source_dir (str or PipelineVariable): The absolute, relative, or S3 URI Path to |
| a directory with any other training source code dependencies aside from the entry |
| point file (default: None). If ``source_dir`` is an S3 URI, it must |
| point to a tar.gz file. The structure within this directory is preserved |
| when training on Amazon SageMaker. If 'git_config' is provided, |
| 'source_dir' should be a relative location to a directory in the Git |
| repo. |
| With the following GitHub repo directory structure: |
| |
| .. code:: |
| |
| |----- README.md |
| |----- src |
| |----- train.py |
| |----- test.py |
| |
| if you need 'train.py' as the entry point and 'test.py' as |
| the training source code, you can assign |
| entry_point='train.py' and source_dir='src'. |
| git_config (dict[str, str]): Git configurations used for cloning |
| files, including ``repo``, ``branch``, ``commit``, |
| ``2FA_enabled``, ``username``, ``password``, and ``token``. The |
| ``repo`` field is required. All other fields are optional. |
| ``repo`` specifies the Git repository where your training script |
| is stored. If you don't provide ``branch``, the default value |
| 'master' is used. If you don't provide ``commit``, the latest |
| commit in the specified branch is used. For example, the following config: |
| |
| .. code:: python |
| |
| git_config = { |
| 'repo': 'https://github.com/aws/sagemaker-python-sdk.git', |
| 'branch': 'test-branch-git-config', |
| 'commit': '329bfcf884482002c05ff7f44f62599ebc9f445a' |
| } |
| |
| results in cloning the repo specified in 'repo', then |
| checking out the 'master' branch, and checking out the specified |
| commit. |
| ``2FA_enabled``, ``username``, ``password``, and ``token`` are |
| used for authentication. For GitHub (or other Git) accounts, set |
| ``2FA_enabled`` to 'True' if two-factor authentication is |
| enabled for the account, otherwise set it to 'False'. If you do |
| not provide a value for ``2FA_enabled``, a default value of |
| 'False' is used. CodeCommit does not support two-factor |
| authentication, so do not provide "2FA_enabled" with CodeCommit |
| repositories. |
| |
| For GitHub and other Git repos, when SSH URLs are provided, it |
| doesn't matter whether 2FA is enabled or disabled. You should |
| either have no passphrase for the SSH key pairs or have the |
| ssh-agent configured so that you will not be prompted for the SSH |
| passphrase when you run the 'git clone' command with SSH URLs. When |
| HTTPS URLs are provided, if 2FA is disabled, then either ``token`` |
| or ``username`` and ``password`` are be used for authentication if provided. |
| ``Token`` is prioritized. If 2FA is enabled, only ``token`` is used |
| for authentication if provided. If required authentication info |
| is not provided, the SageMaker Python SDK attempts to use local credentials |
| to authenticate. If that fails, an error message is thrown. |
| |
| For CodeCommit repos, 2FA is not supported, so '2FA_enabled' |
| should not be provided. There is no token in CodeCommit, so |
| ``token`` should also not be provided. When ``repo`` is an SSH URL, |
| the requirements are the same as GitHub repos. When ``repo`` |
| is an HTTPS URL, ``username`` and ``password`` are used for |
| authentication if they are provided. If they are not provided, |
| the SageMaker Python SDK attempts to use either the CodeCommit |
| credential helper or local credential storage for authentication. |
| hyperparameters (dict[str, str] or dict[str, PipelineVariable]): |
| A dictionary containing the hyperparameters to |
| initialize this estimator with. (Default: None). |
| container_log_level (int or PipelineVariable): The log level to use within the container |
| (default: logging.INFO). Valid values are defined in the Python |
| logging module. |
| code_location (str): The S3 prefix URI where custom code is |
| uploaded (default: None). You must not include a trailing slash because |
| a string prepended with a "/" is appended to ``code_location``. The code |
| file uploaded to S3 is 'code_location/job-name/source/sourcedir.tar.gz'. |
| If not specified, the default ``code location`` is 's3://output_bucket/job-name/'. |
| entry_point (str or PipelineVariable): The absolute or relative path to the local Python |
| source file that should be executed as the entry point to |
| training. (Default: None). If ``source_dir`` is specified, then ``entry_point`` |
| must point to a file located at the root of ``source_dir``. |
| If 'git_config' is provided, 'entry_point' should be |
| a relative location to the Python source file in the Git repo. |
| |
| Example: |
| With the following GitHub repo directory structure: |
| |
| >>> |----- README.md |
| >>> |----- src |
| >>> |----- train.py |
| >>> |----- test.py |
| |
| You can assign entry_point='src/train.py'. |
| dependencies (list[str]): A list of absolute or relative paths to directories |
| with any additional libraries that should be exported |
| to the container (default: []). The library folders are |
| copied to SageMaker in the same folder where the entrypoint is |
| copied. If 'git_config' is provided, 'dependencies' should be a |
| list of relative locations to directories with any additional |
| libraries needed in the Git repo. |
| |
| .. admonition:: Example |
| |
| The following Estimator call: |
| |
| >>> Estimator(entry_point='train.py', |
| ... dependencies=['my/libs/common', 'virtual-env']) |
| |
| results in the following structure inside the container: |
| |
| >>> $ ls |
| |
| >>> opt/ml/code |
| >>> |------ train.py |
| >>> |------ common |
| >>> |------ virtual-env |
| |
| This is not supported with "local code" in Local Mode. |
| instance_groups (list[:class:`sagemaker.instance_group.InstanceGroup`]): |
| Optional. A list of ``InstanceGroup`` objects |
| for launching a training job with a heterogeneous cluster. |
| For example: |
| |
| .. code:: python |
| |
| instance_groups=[ |
| sagemaker.InstanceGroup( |
| 'instance_group_name_1', 'ml.p3dn.24xlarge', 64), |
| sagemaker.InstanceGroup( |
| 'instance_group_name_2', 'ml.c5n.18xlarge', 64)] |
| |
| For instructions on how to use ``InstanceGroup`` objects |
| to configure a heterogeneous cluster |
| through the SageMaker generic and framework estimator classes, see |
| `Train Using a Heterogeneous Cluster |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/train-heterogeneous-cluster.html>`_ |
| in the *Amazon SageMaker developer guide*. |
| """ |
| instance_count = renamed_kwargs( |
| "train_instance_count", "instance_count", instance_count, kwargs |
| ) |
| instance_type = renamed_kwargs( |
| "train_instance_type", "instance_type", instance_type, kwargs |
| ) |
| max_run = renamed_kwargs("train_max_run", "max_run", max_run, kwargs) |
| use_spot_instances = renamed_kwargs( |
| "train_use_spot_instances", "use_spot_instances", use_spot_instances, kwargs |
| ) |
| max_wait = renamed_kwargs("train_max_wait", "max_wait", max_wait, kwargs) |
| volume_size = renamed_kwargs("train_volume_size", "volume_size", volume_size, kwargs) |
| volume_kms_key = renamed_kwargs( |
| "train_volume_kms_key", "volume_kms_key", volume_kms_key, kwargs |
| ) |
|
|
| validate_source_code_input_against_pipeline_variables( |
| entry_point=entry_point, |
| source_dir=source_dir, |
| git_config=git_config, |
| enable_network_isolation=enable_network_isolation, |
| ) |
|
|
| self.role = role |
| self.instance_count = instance_count |
| self.instance_type = instance_type |
| self.keep_alive_period_in_seconds = keep_alive_period_in_seconds |
| self.instance_groups = instance_groups |
| self.volume_size = volume_size |
| self.volume_kms_key = volume_kms_key |
| self.max_run = max_run |
| self.input_mode = input_mode |
| self.metric_definitions = metric_definitions |
| self.model_uri = model_uri |
| self.model_channel_name = model_channel_name |
| self.code_uri = None |
| self.code_channel_name = "code" |
| self.source_dir = source_dir |
| self.git_config = git_config |
| self.container_log_level = container_log_level |
| self._hyperparameters = hyperparameters.copy() if hyperparameters else {} |
| self.code_location = code_location |
| self.entry_point = entry_point |
| self.dependencies = dependencies or [] |
| self.uploaded_code = None |
| self.tags = add_jumpstart_tags( |
| tags=tags, training_model_uri=self.model_uri, training_script_uri=self.source_dir |
| ) |
| if self.instance_type in ("local", "local_gpu"): |
| if self.instance_type == "local_gpu" and self.instance_count > 1: |
| raise RuntimeError("Distributed Training in Local GPU is not supported") |
| self.sagemaker_session = sagemaker_session or LocalSession() |
| if not isinstance(self.sagemaker_session, sagemaker.local.LocalSession): |
| raise RuntimeError( |
| "instance_type local or local_gpu is only supported with an" |
| "instance of LocalSession" |
| ) |
| else: |
| self.sagemaker_session = sagemaker_session or Session() |
|
|
| self.base_job_name = base_job_name |
| self._current_job_name = None |
| if ( |
| not self.sagemaker_session.local_mode |
| and output_path |
| and not is_pipeline_variable(output_path) |
| and output_path.startswith("file://") |
| ): |
| raise RuntimeError("file:// output paths are only supported in Local Mode") |
| self.output_path = output_path |
| self.output_kms_key = output_kms_key |
| self.latest_training_job = None |
| self.jobs = [] |
| self.deploy_instance_type = None |
|
|
| self._compiled_models = {} |
|
|
| |
| self.subnets = subnets |
| self.security_group_ids = security_group_ids |
|
|
| self.encrypt_inter_container_traffic = encrypt_inter_container_traffic |
| self.use_spot_instances = use_spot_instances |
| self.max_wait = max_wait |
| self.checkpoint_s3_uri = checkpoint_s3_uri |
| self.checkpoint_local_path = checkpoint_local_path |
|
|
| self.rules = rules |
| self.debugger_hook_config = debugger_hook_config |
| self.tensorboard_output_config = tensorboard_output_config |
|
|
| self.debugger_rule_configs = None |
| self.collection_configs = None |
|
|
| self.enable_sagemaker_metrics = enable_sagemaker_metrics |
| self._enable_network_isolation = enable_network_isolation |
|
|
| self.profiler_config = profiler_config |
| self.disable_profiler = disable_profiler |
|
|
| self.environment = environment |
|
|
| self.max_retry_attempts = max_retry_attempts |
|
|
| if not _region_supports_profiler(self.sagemaker_session.boto_region_name): |
| self.disable_profiler = True |
|
|
| self.profiler_rule_configs = None |
| self.profiler_rules = None |
| self.debugger_rules = None |
|
|
| @abstractmethod |
| def training_image_uri(self): |
| """Return the Docker image to use for training. |
| |
| The :meth:`~sagemaker.estimator.EstimatorBase.fit` method, which does |
| the model training, calls this method to find the image to use for model |
| training. |
| |
| Returns: |
| str: The URI of the Docker image. |
| """ |
|
|
| @abstractmethod |
| def hyperparameters(self): |
| """Return the hyperparameters as a dictionary to use for training. |
| |
| The :meth:`~sagemaker.estimator.EstimatorBase.fit` method, which |
| trains the model, calls this method to find the hyperparameters. |
| |
| Returns: |
| dict[str, str]: The hyperparameters. |
| """ |
|
|
| def enable_network_isolation(self): |
| """Return True if this Estimator will need network isolation to run. |
| |
| Returns: |
| bool: Whether this Estimator needs network isolation or not. |
| """ |
| return self._enable_network_isolation |
|
|
| def prepare_workflow_for_training(self, job_name=None): |
| """Calls _prepare_for_training. Used when setting up a workflow. |
| |
| Args: |
| job_name (str): Name of the training job to be created. If not |
| specified, one is generated, using the base name given to the |
| constructor if applicable. |
| """ |
| self._prepare_for_training(job_name=job_name) |
|
|
| def _ensure_base_job_name(self): |
| """Set ``self.base_job_name`` if it is not set already.""" |
| |
| self.base_job_name = ( |
| self.base_job_name |
| or get_jumpstart_base_name_if_jumpstart_model(self.source_dir, self.model_uri) |
| or base_name_from_image( |
| self.training_image_uri(), default_base_name=EstimatorBase.JOB_CLASS_NAME |
| ) |
| ) |
|
|
| def _get_or_create_name(self, name=None): |
| """Generate a name based on the base job name or training image if needed. |
| |
| Args: |
| name (str): User-supplied name. If not specified, a name is generated from |
| the base job name or training image. |
| |
| Returns: |
| str: Either the user-supplied name or a generated name. |
| """ |
| if name: |
| return name |
|
|
| self._ensure_base_job_name() |
| return name_from_base(self.base_job_name) |
|
|
| @staticmethod |
| def _json_encode_hyperparameters(hyperparameters: Dict[str, Any]) -> Dict[str, Any]: |
| """Applies JSON encoding for certain hyperparameter types, returns hyperparameters. |
| |
| Args: |
| hyperparameters (dict): Dictionary of hyperparameters. |
| """ |
| current_hyperparameters = hyperparameters |
| if current_hyperparameters is not None: |
| hyperparameters = { |
| str(k): (v.to_string() if is_pipeline_variable(v) else json.dumps(v)) |
| for (k, v) in current_hyperparameters.items() |
| } |
| return hyperparameters |
|
|
| def _prepare_for_training(self, job_name=None): |
| """Set any values in the estimator that need to be set before training. |
| |
| Args: |
| job_name (str): Name of the training job to be created. If not |
| specified, one is generated, using the base name given to the |
| constructor if applicable. |
| """ |
| self._current_job_name = self._get_or_create_name(job_name) |
|
|
| |
| |
| if self.output_path is None: |
| local_code = get_config_value("local.local_code", self.sagemaker_session.config) |
| if self.sagemaker_session.local_mode and local_code: |
| self.output_path = "" |
| else: |
| self.output_path = "s3://{}/".format(self.sagemaker_session.default_bucket()) |
|
|
| if self.git_config: |
| updated_paths = git_utils.git_clone_repo( |
| self.git_config, self.entry_point, self.source_dir, self.dependencies |
| ) |
| self.entry_point = updated_paths["entry_point"] |
| self.source_dir = updated_paths["source_dir"] |
| self.dependencies = updated_paths["dependencies"] |
|
|
| if self.source_dir or self.entry_point or self.dependencies: |
|
|
| |
| |
| |
| if ( |
| self.source_dir |
| and not is_pipeline_variable(self.source_dir) |
| and not self.source_dir.lower().startswith("s3://") |
| ): |
| validate_source_dir(self.entry_point, self.source_dir) |
|
|
| |
| |
| local_code = get_config_value("local.local_code", self.sagemaker_session.config) |
|
|
| if self.sagemaker_session.local_mode and local_code: |
| |
| if self.source_dir is None: |
| self.source_dir = os.path.dirname(self.entry_point) |
| self.entry_point = os.path.basename(self.entry_point) |
|
|
| code_dir = "file://" + self.source_dir |
| script = self.entry_point |
| elif self.enable_network_isolation() and self.entry_point: |
| self.uploaded_code = self._stage_user_code_in_s3() |
| code_dir = self.CONTAINER_CODE_CHANNEL_SOURCEDIR_PATH |
| script = self.uploaded_code.script_name |
| self.code_uri = self.uploaded_code.s3_prefix |
| else: |
| self.uploaded_code = self._stage_user_code_in_s3() |
| code_dir = self.uploaded_code.s3_prefix |
| script = self.uploaded_code.script_name |
|
|
| |
| |
| self._script_mode_hyperparam_update(code_dir, script) |
|
|
| self._prepare_rules() |
| self._prepare_debugger_for_training() |
| self._prepare_profiler_for_training() |
|
|
| def _script_mode_hyperparam_update(self, code_dir: str, script: str) -> None: |
| """Applies in-place updates to hyperparameters required for script mode with training. |
| |
| Args: |
| code_dir (str): The directory hosting the training scripts. |
| script (str): The relative filepath of the training entry-point script. |
| """ |
| hyperparams: Dict[str, str] = {} |
| hyperparams[DIR_PARAM_NAME] = code_dir |
| hyperparams[SCRIPT_PARAM_NAME] = script |
| hyperparams[CONTAINER_LOG_LEVEL_PARAM_NAME] = self.container_log_level |
| hyperparams[JOB_NAME_PARAM_NAME] = self._current_job_name |
| hyperparams[SAGEMAKER_REGION_PARAM_NAME] = self.sagemaker_session.boto_region_name |
|
|
| self._hyperparameters.update(EstimatorBase._json_encode_hyperparameters(hyperparams)) |
|
|
| def _stage_user_code_in_s3(self) -> str: |
| """Uploads the user training script to S3 and returns the S3 URI. |
| |
| Returns: S3 URI |
| """ |
| if is_pipeline_variable(self.output_path): |
| if self.code_location is None: |
| code_bucket = self.sagemaker_session.default_bucket() |
| code_s3_prefix = "{}/{}".format(self._current_job_name, "source") |
| kms_key = None |
| else: |
| code_bucket, key_prefix = parse_s3_url(self.code_location) |
| code_s3_prefix = "/".join( |
| filter(None, [key_prefix, self._current_job_name, "source"]) |
| ) |
|
|
| output_bucket = self.sagemaker_session.default_bucket() |
| kms_key = self.output_kms_key if code_bucket == output_bucket else None |
| else: |
| local_mode = self.output_path.startswith("file://") |
| if local_mode: |
| if self.code_location is None: |
| code_bucket = self.sagemaker_session.default_bucket() |
| code_s3_prefix = "{}/{}".format(self._current_job_name, "source") |
| kms_key = None |
| else: |
| code_bucket, key_prefix = parse_s3_url(self.code_location) |
| code_s3_prefix = "/".join( |
| filter(None, [key_prefix, self._current_job_name, "source"]) |
| ) |
| kms_key = None |
| else: |
| if self.code_location is None: |
| code_bucket, _ = parse_s3_url(self.output_path) |
| code_s3_prefix = "{}/{}".format(self._current_job_name, "source") |
| kms_key = self.output_kms_key |
| else: |
| code_bucket, key_prefix = parse_s3_url(self.code_location) |
| code_s3_prefix = "/".join( |
| filter(None, [key_prefix, self._current_job_name, "source"]) |
| ) |
|
|
| output_bucket, _ = parse_s3_url(self.output_path) |
| kms_key = self.output_kms_key if code_bucket == output_bucket else None |
|
|
| return tar_and_upload_dir( |
| session=self.sagemaker_session.boto_session, |
| bucket=code_bucket, |
| s3_key_prefix=code_s3_prefix, |
| script=self.entry_point, |
| directory=self.source_dir, |
| dependencies=self.dependencies, |
| kms_key=kms_key, |
| s3_resource=self.sagemaker_session.s3_resource, |
| settings=self.sagemaker_session.settings, |
| ) |
|
|
| def _prepare_rules(self): |
| """Rules list includes both debugger and profiler rules. |
| |
| Customer can explicitly disable any rule by setting rules to an empty list. |
| """ |
| self.debugger_rules = [] |
| self.profiler_rules = [] |
| if self.rules is not None: |
| for rule in self.rules: |
| if isinstance(rule, Rule): |
| |
| self._check_debugger_rule(rule) |
| self.debugger_rules.append(rule) |
| elif isinstance(rule, ProfilerRule): |
| self.profiler_rules.append(rule) |
| else: |
| raise RuntimeError( |
| "Rules list can only contain sagemaker.debugger.Rule " |
| + "and sagemaker.debugger.ProfilerRule" |
| ) |
|
|
| def _check_debugger_rule(self, rule): |
| """Add warning for incorrectly used xgboost rules.""" |
| _xgboost_specific_rules = ["FeatureImportanceOverweight", "TreeDepth"] |
| if rule.name in _xgboost_specific_rules: |
| logger.warning( |
| "TreeDepth and FeatureImportanceOverweight rules are valid " |
| "only for the XGBoost algorithm. Please make sure this estimator " |
| "is used for XGBoost algorithm. " |
| ) |
|
|
| def _prepare_debugger_for_training(self): |
| """Prepare debugger rules and debugger configs for training.""" |
| if self.debugger_rules and self.debugger_hook_config is None: |
| self.debugger_hook_config = DebuggerHookConfig(s3_output_path=self.output_path) |
| |
| if self.debugger_hook_config and not self.debugger_hook_config.s3_output_path: |
| self.debugger_hook_config.s3_output_path = self.output_path |
| self.debugger_rule_configs = self._prepare_debugger_rules() |
| self._prepare_collection_configs() |
| self._validate_and_set_debugger_configs() |
| if not self.debugger_hook_config: |
| if self.environment is None: |
| self.environment = {} |
| self.environment[DEBUGGER_FLAG] = "0" |
|
|
| def _validate_and_set_debugger_configs(self): |
| """Set defaults for debugging.""" |
| region_supports_debugger = _region_supports_debugger( |
| self.sagemaker_session.boto_region_name |
| ) |
|
|
| if region_supports_debugger: |
| if self.debugger_hook_config in [None, {}]: |
| self.debugger_hook_config = DebuggerHookConfig(s3_output_path=self.output_path) |
| else: |
| if self.debugger_hook_config is not False and self.debugger_hook_config: |
| |
| raise ValueError( |
| "Current region does not support debugger but debugger hook config is set!" |
| ) |
| |
| self.debugger_hook_config = False |
|
|
| def _prepare_debugger_rules(self): |
| """Set any necessary values in debugger rules, if they are provided.""" |
| debugger_rule_configs = [] |
| if self.debugger_rules: |
| for rule in self.debugger_rules: |
| self._set_default_rule_config(rule) |
| self._set_source_s3_uri(rule) |
| rule.prepare_actions(self._current_job_name) |
| debugger_rule_configs.append(rule.to_debugger_rule_config_dict()) |
| return debugger_rule_configs |
|
|
| def _prepare_collection_configs(self): |
| """De-duplicate configurations and save them in the debugger hook configuration.""" |
| |
| self.collection_configs = set() |
| |
| if self.debugger_rules: |
| for rule in self.debugger_rules: |
| self.collection_configs.update(rule.collection_configs) |
| |
| if self.debugger_hook_config: |
| self.collection_configs.update(self.debugger_hook_config.collection_configs or []) |
|
|
| def _prepare_profiler_for_training(self): |
| """Set necessary values and do basic validations in profiler config and profiler rules. |
| |
| When user explicitly set rules to an empty list, default profiler rule won't be enabled. |
| Default profiler rule will be enabled in supported regions when either: |
| 1. user doesn't specify any rules, i.e., rules=None; or |
| 2. user only specify debugger rules, i.e., rules=[Rule.sagemaker(...)] |
| """ |
| if self.disable_profiler: |
| if self.profiler_config: |
| raise RuntimeError("profiler_config cannot be set when disable_profiler is True.") |
| if self.profiler_rules: |
| raise RuntimeError("ProfilerRule cannot be set when disable_profiler is True.") |
| elif _region_supports_profiler(self.sagemaker_session.boto_region_name): |
| if self.profiler_config is None: |
| self.profiler_config = ProfilerConfig(s3_output_path=self.output_path) |
| if self.rules is None or (self.rules and not self.profiler_rules): |
| self.profiler_rules = [get_default_profiler_rule()] |
|
|
| if self.profiler_config and not self.profiler_config.s3_output_path: |
| self.profiler_config.s3_output_path = self.output_path |
|
|
| self.profiler_rule_configs = self._prepare_profiler_rules() |
|
|
| def _prepare_profiler_rules(self): |
| """Set any necessary values in profiler rules, if they are provided.""" |
| profiler_rule_configs = [] |
| if self.profiler_rules: |
| for rule in self.profiler_rules: |
| self._set_default_rule_config(rule) |
| self._set_source_s3_uri(rule) |
| profiler_rule_configs.append(rule.to_profiler_rule_config_dict()) |
| return profiler_rule_configs |
|
|
| def _set_default_rule_config(self, rule): |
| """Set default rule configurations. |
| |
| Args: |
| rule (:class:`~sagemaker.debugger.RuleBase`): Any rule object that derives from RuleBase |
| """ |
| if rule.image_uri == "DEFAULT_RULE_EVALUATOR_IMAGE": |
| rule.image_uri = get_rule_container_image_uri(self.sagemaker_session.boto_region_name) |
| rule.instance_type = None |
| rule.volume_size_in_gb = None |
|
|
| def _set_source_s3_uri(self, rule): |
| """Set updated source S3 uri when specified. |
| |
| Args: |
| rule (:class:`~sagemaker.debugger.RuleBase`): Any rule object that derives from RuleBase |
| """ |
| if "source_s3_uri" in (rule.rule_parameters or {}): |
| parse_result = urlparse(rule.rule_parameters["source_s3_uri"]) |
| if parse_result.scheme != "s3": |
| desired_s3_uri = os.path.join( |
| "s3://", |
| self.sagemaker_session.default_bucket(), |
| rule.name, |
| str(uuid.uuid4()), |
| ) |
| s3_uri = S3Uploader.upload( |
| local_path=rule.rule_parameters["source_s3_uri"], |
| desired_s3_uri=desired_s3_uri, |
| sagemaker_session=self.sagemaker_session, |
| ) |
| rule.rule_parameters["source_s3_uri"] = s3_uri |
|
|
| def latest_job_debugger_artifacts_path(self): |
| """Gets the path to the DebuggerHookConfig output artifacts. |
| |
| Returns: |
| str: An S3 path to the output artifacts. |
| """ |
| self._ensure_latest_training_job( |
| error_message="""Cannot get the Debugger artifacts path. |
| The Estimator is not associated with a training job.""" |
| ) |
| if self.debugger_hook_config is not None: |
| return os.path.join( |
| self.debugger_hook_config.s3_output_path, |
| self.latest_training_job.name, |
| "debug-output", |
| ) |
| return None |
|
|
| def latest_job_tensorboard_artifacts_path(self): |
| """Gets the path to the TensorBoardOutputConfig output artifacts. |
| |
| Returns: |
| str: An S3 path to the output artifacts. |
| """ |
| self._ensure_latest_training_job( |
| error_message="""Cannot get the TensorBoard artifacts path. |
| The Estimator is not associated with a training job.""" |
| ) |
| if self.debugger_hook_config is not None: |
| return os.path.join( |
| self.tensorboard_output_config.s3_output_path, |
| self.latest_training_job.name, |
| "tensorboard-output", |
| ) |
| return None |
|
|
| def latest_job_profiler_artifacts_path(self): |
| """Gets the path to the profiling output artifacts. |
| |
| Returns: |
| str: An S3 path to the output artifacts. |
| """ |
| self._ensure_latest_training_job( |
| error_message="""Cannot get the profiling output artifacts path. |
| The Estimator is not associated with a training job.""" |
| ) |
| if self.profiler_config is not None: |
| return os.path.join( |
| self.profiler_config.s3_output_path, |
| self.latest_training_job.name, |
| "profiler-output", |
| ) |
| return None |
|
|
| @runnable_by_pipeline |
| def fit( |
| self, |
| inputs: Optional[Union[str, Dict, TrainingInput, FileSystemInput]] = None, |
| wait: bool = True, |
| logs: str = "All", |
| job_name: Optional[str] = None, |
| experiment_config: Optional[Dict[str, str]] = None, |
| ): |
| """Train a model using the input training dataset. |
| |
| The API calls the Amazon SageMaker CreateTrainingJob API to start |
| model training. The API uses configuration you provided to create the |
| estimator and the specified input training data to send the |
| CreatingTrainingJob request to Amazon SageMaker. |
| |
| This is a synchronous operation. After the model training |
| successfully completes, you can call the ``deploy()`` method to host the |
| model using the Amazon SageMaker hosting services. |
| |
| Args: |
| inputs (str or dict or sagemaker.inputs.TrainingInput or |
| sagemaker.inputs.FileSystemInput): Information about the training data. |
| This can be one of four types: |
| |
| * (str) the S3 location where training data is saved, or a file:// path in |
| local mode. |
| * (dict[str, str] or dict[str, sagemaker.inputs.TrainingInput] or |
| dict[str, sagemaker.inputs.FileSystemInput]) If using multiple channels for |
| training data, you can specify a dict mapping channel names to strings or |
| :func:`~sagemaker.inputs.TrainingInput` objects or |
| :func:`~sagemaker.inputs.FileSystemInput` objects. |
| * (sagemaker.inputs.TrainingInput) - channel configuration for S3 data sources |
| that can provide additional information as well as the path to the training |
| dataset. |
| See :func:`sagemaker.inputs.TrainingInput` for full details. |
| * (sagemaker.inputs.FileSystemInput) - channel configuration for |
| a file system data source that can provide additional information as well as |
| the path to the training dataset. |
| |
| wait (bool): Whether the call should wait until the job completes (default: True). |
| logs ([str]): A list of strings specifying which logs to print. Acceptable |
| strings are "All", "None", "Training", or "Rules". To maintain backwards |
| compatibility, boolean values are also accepted and converted to strings. |
| Only meaningful when wait is True. |
| job_name (str): Training job name. If not specified, the estimator generates |
| a default job name based on the training image name and current timestamp. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| * Both `ExperimentName` and `TrialName` will be ignored if the Estimator instance |
| is built with :class:`~sagemaker.workflow.pipeline_context.PipelineSession`. |
| However, the value of `TrialComponentDisplayName` is honored for display in Studio. |
| Returns: |
| None or pipeline step arguments in case the Estimator instance is built with |
| :class:`~sagemaker.workflow.pipeline_context.PipelineSession` |
| """ |
| self._prepare_for_training(job_name=job_name) |
|
|
| self.latest_training_job = _TrainingJob.start_new(self, inputs, experiment_config) |
| self.jobs.append(self.latest_training_job) |
| if wait: |
| self.latest_training_job.wait(logs=logs) |
|
|
| def _compilation_job_name(self): |
| """Placeholder docstring""" |
| base_name = self.base_job_name or base_name_from_image( |
| self.training_image_uri(), default_base_name=EstimatorBase.JOB_CLASS_NAME |
| ) |
| return name_from_base("compilation-" + base_name) |
|
|
| def compile_model( |
| self, |
| target_instance_family, |
| input_shape, |
| output_path, |
| framework=None, |
| framework_version=None, |
| compile_max_run=15 * 60, |
| tags=None, |
| target_platform_os=None, |
| target_platform_arch=None, |
| target_platform_accelerator=None, |
| compiler_options=None, |
| **kwargs, |
| ): |
| """Compile a Neo model using the input model. |
| |
| Args: |
| target_instance_family (str): Identifies the device that you want to |
| run your model after compilation, for example: ml_c5. For allowed |
| strings see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_OutputConfig.html. |
| input_shape (dict): Specifies the name and shape of the expected |
| inputs for your trained model in json dictionary form, for |
| example: {'data':[1,3,1024,1024]}, or {'var1': [1,1,28,28], |
| 'var2':[1,1,28,28]} |
| output_path (str): Specifies where to store the compiled model |
| framework (str): The framework that is used to train the original |
| model. Allowed values: 'mxnet', 'tensorflow', 'keras', 'pytorch', |
| 'onnx', 'xgboost' |
| framework_version (str): The version of the framework |
| compile_max_run (int): Timeout in seconds for compilation (default: |
| 15 * 60). After this amount of time Amazon SageMaker Neo |
| terminates the compilation job regardless of its current status. |
| tags (list[dict]): List of tags for labeling a compilation job. For |
| more, see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_Tag.html. |
| target_platform_os (str): Target Platform OS, for example: 'LINUX'. |
| For allowed strings see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_OutputConfig.html. |
| It can be used instead of target_instance_family. |
| target_platform_arch (str): Target Platform Architecture, for example: 'X86_64'. |
| For allowed strings see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_OutputConfig.html. |
| It can be used instead of target_instance_family. |
| target_platform_accelerator (str, optional): Target Platform Accelerator, |
| for example: 'NVIDIA'. For allowed strings see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_OutputConfig.html. |
| It can be used instead of target_instance_family. |
| compiler_options (dict, optional): Additional parameters for compiler. |
| Compiler Options are TargetPlatform / target_instance_family specific. See |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_OutputConfig.html for details. |
| **kwargs: Passed to invocation of ``create_model()``. |
| Implementations may customize ``create_model()`` to accept |
| ``**kwargs`` to customize model creation during deploy. For |
| more, see the implementation docs. |
| |
| Returns: |
| sagemaker.model.Model: A SageMaker ``Model`` object. See |
| :func:`~sagemaker.model.Model` for full details. |
| """ |
| if framework and framework not in NEO_ALLOWED_FRAMEWORKS: |
| raise ValueError( |
| "Please use valid framework, allowed values: {}".format(NEO_ALLOWED_FRAMEWORKS) |
| ) |
|
|
| if (framework is None) != (framework_version is None): |
| raise ValueError("You should provide framework and framework_version at the same time.") |
|
|
| model = self.create_model(**kwargs) |
|
|
| self._compiled_models[target_instance_family] = model.compile( |
| target_instance_family, |
| input_shape, |
| output_path, |
| self.role, |
| tags, |
| self._compilation_job_name(), |
| compile_max_run, |
| framework=framework, |
| framework_version=framework_version, |
| target_platform_os=target_platform_os, |
| target_platform_arch=target_platform_arch, |
| target_platform_accelerator=target_platform_accelerator, |
| compiler_options=compiler_options, |
| ) |
| return self._compiled_models[target_instance_family] |
|
|
| @classmethod |
| def attach(cls, training_job_name, sagemaker_session=None, model_channel_name="model"): |
| """Attach to an existing training job. |
| |
| Create an Estimator bound to an existing training job, each subclass |
| is responsible to implement |
| ``_prepare_init_params_from_job_description()`` as this method delegates |
| the actual conversion of a training job description to the arguments |
| that the class constructor expects. After attaching, if the training job |
| has a Complete status, it can be ``deploy()`` ed to create a SageMaker |
| Endpoint and return a ``Predictor``. |
| |
| If the training job is in progress, attach will block until the training job |
| completes, but logs of the training job will not display. To see the logs |
| content, please call ``logs()`` |
| |
| Examples: |
| >>> my_estimator.fit(wait=False) |
| >>> training_job_name = my_estimator.latest_training_job.name |
| Later on: |
| >>> attached_estimator = Estimator.attach(training_job_name) |
| >>> attached_estimator.logs() |
| >>> attached_estimator.deploy() |
| |
| Args: |
| training_job_name (str): The name of the training job to attach to. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the estimator creates one |
| using the default AWS configuration chain. |
| model_channel_name (str): Name of the channel where pre-trained |
| model data will be downloaded (default: 'model'). If no channel |
| with the same name exists in the training job, this option will |
| be ignored. |
| |
| Returns: |
| Instance of the calling ``Estimator`` Class with the attached |
| training job. |
| """ |
| sagemaker_session = sagemaker_session or Session() |
|
|
| job_details = sagemaker_session.sagemaker_client.describe_training_job( |
| TrainingJobName=training_job_name |
| ) |
| init_params = cls._prepare_init_params_from_job_description(job_details, model_channel_name) |
| tags = sagemaker_session.sagemaker_client.list_tags( |
| ResourceArn=job_details["TrainingJobArn"] |
| )["Tags"] |
| init_params.update(tags=tags) |
|
|
| estimator = cls(sagemaker_session=sagemaker_session, **init_params) |
| estimator.latest_training_job = _TrainingJob( |
| sagemaker_session=sagemaker_session, job_name=training_job_name |
| ) |
| estimator._current_job_name = estimator.latest_training_job.name |
| estimator.latest_training_job.wait(logs="None") |
| return estimator |
|
|
| def logs(self): |
| """Display the logs for Estimator's training job. |
| |
| If the output is a tty or a Jupyter cell, it will be color-coded based |
| on which instance the log entry is from. |
| """ |
| self.sagemaker_session.logs_for_job(self.latest_training_job.name, wait=True) |
|
|
| def deploy( |
| self, |
| initial_instance_count=None, |
| instance_type=None, |
| serializer=None, |
| deserializer=None, |
| accelerator_type=None, |
| endpoint_name=None, |
| use_compiled_model=False, |
| wait=True, |
| model_name=None, |
| kms_key=None, |
| data_capture_config=None, |
| tags=None, |
| serverless_inference_config=None, |
| async_inference_config=None, |
| **kwargs, |
| ): |
| """Deploy the trained model to an Amazon SageMaker endpoint. |
| |
| And then return ``sagemaker.Predictor`` object. |
| |
| More information: |
| http://docs.aws.amazon.com/sagemaker/latest/dg/how-it-works-training.html |
| |
| Args: |
| initial_instance_count (int): The initial number of instances to run |
| in the ``Endpoint`` created from this ``Model``. If not using |
| serverless inference, then it need to be a number larger or equals |
| to 1 (default: None) |
| instance_type (str): The EC2 instance type to deploy this Model to. |
| For example, 'ml.p2.xlarge', or 'local' for local mode. If not using |
| serverless inference, then it is required to deploy a model. |
| (default: None) |
| serializer (:class:`~sagemaker.serializers.BaseSerializer`): A |
| serializer object, used to encode data for an inference endpoint |
| (default: None). If ``serializer`` is not None, then |
| ``serializer`` will override the default serializer. The |
| default serializer is set by the ``predictor_cls``. |
| deserializer (:class:`~sagemaker.deserializers.BaseDeserializer`): A |
| deserializer object, used to decode data from an inference |
| endpoint (default: None). If ``deserializer`` is not None, then |
| ``deserializer`` will override the default deserializer. The |
| default deserializer is set by the ``predictor_cls``. |
| accelerator_type (str): Type of Elastic Inference accelerator to |
| attach to an endpoint for model loading and inference, for |
| example, 'ml.eia1.medium'. If not specified, no Elastic |
| Inference accelerator will be attached to the endpoint. For more |
| information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/ei.html |
| endpoint_name (str): Name to use for creating an Amazon SageMaker |
| endpoint. If not specified, the name of the training job is |
| used. |
| use_compiled_model (bool): Flag to select whether to use compiled |
| (optimized) model. Default: False. |
| wait (bool): Whether the call should wait until the deployment of |
| model completes (default: True). |
| model_name (str): Name to use for creating an Amazon SageMaker |
| model. If not specified, the estimator generates a default job name |
| based on the training image name and current timestamp. |
| kms_key (str): The ARN of the KMS key that is used to encrypt the |
| data on the storage volume attached to the instance hosting the |
| endpoint. |
| data_capture_config (sagemaker.model_monitor.DataCaptureConfig): Specifies |
| configuration related to Endpoint data capture for use with |
| Amazon SageMaker Model Monitoring. Default: None. |
| async_inference_config (sagemaker.model_monitor.AsyncInferenceConfig): Specifies |
| configuration related to async inference. Use this configuration when trying |
| to create async endpoint and make async inference. If empty config object |
| passed through, will use default config to deploy async endpoint. Deploy a |
| real-time endpoint if it's None. (default: None) |
| serverless_inference_config (sagemaker.serverless.ServerlessInferenceConfig): |
| Specifies configuration related to serverless endpoint. Use this configuration |
| when trying to create serverless endpoint and make serverless inference. If |
| empty object passed through, will use pre-defined values in |
| ``ServerlessInferenceConfig`` class to deploy serverless endpoint. Deploy an |
| instance based endpoint if it's None. (default: None) |
| tags(List[dict[str, str]]): Optional. The list of tags to attach to this specific |
| endpoint. Example: |
| >>> tags = [{'Key': 'tagname', 'Value': 'tagvalue'}] |
| For more information about tags, see |
| https://boto3.amazonaws.com/v1/documentation\ |
| /api/latest/reference/services/sagemaker.html#SageMaker.Client.add_tags |
| **kwargs: Passed to invocation of ``create_model()``. |
| Implementations may customize ``create_model()`` to accept |
| ``**kwargs`` to customize model creation during deploy. |
| For more, see the implementation docs. |
| |
| Returns: |
| sagemaker.predictor.Predictor: A predictor that provides a ``predict()`` method, |
| which can be used to send requests to the Amazon SageMaker |
| endpoint and obtain inferences. |
| """ |
| removed_kwargs("update_endpoint", kwargs) |
| is_serverless = serverless_inference_config is not None |
| self._ensure_latest_training_job() |
| self._ensure_base_job_name() |
|
|
| jumpstart_base_name = get_jumpstart_base_name_if_jumpstart_model( |
| kwargs.get("source_dir"), self.source_dir, kwargs.get("model_data"), self.model_uri |
| ) |
| default_name = ( |
| name_from_base(jumpstart_base_name) |
| if jumpstart_base_name |
| else name_from_base(self.base_job_name) |
| ) |
| endpoint_name = endpoint_name or default_name |
| model_name = model_name or default_name |
|
|
| self.deploy_instance_type = instance_type |
| if use_compiled_model and not is_serverless: |
| family = "_".join(instance_type.split(".")[:-1]) |
| if family not in self._compiled_models: |
| raise ValueError( |
| "No compiled model for {}. " |
| "Please compile one with compile_model before deploying.".format(family) |
| ) |
| model = self._compiled_models[family] |
| else: |
| kwargs["model_kms_key"] = self.output_kms_key |
| model = self.create_model(**kwargs) |
|
|
| model.name = model_name |
|
|
| tags = update_inference_tags_with_jumpstart_training_tags( |
| inference_tags=tags, training_tags=self.tags |
| ) |
|
|
| return model.deploy( |
| instance_type=instance_type, |
| initial_instance_count=initial_instance_count, |
| serializer=serializer, |
| deserializer=deserializer, |
| accelerator_type=accelerator_type, |
| endpoint_name=endpoint_name, |
| tags=tags or self.tags, |
| wait=wait, |
| kms_key=kms_key, |
| data_capture_config=data_capture_config, |
| serverless_inference_config=serverless_inference_config, |
| async_inference_config=async_inference_config, |
| ) |
|
|
| def register( |
| self, |
| content_types, |
| response_types, |
| inference_instances=None, |
| transform_instances=None, |
| image_uri=None, |
| model_package_name=None, |
| model_package_group_name=None, |
| model_metrics=None, |
| metadata_properties=None, |
| marketplace_cert=False, |
| approval_status=None, |
| description=None, |
| compile_model_family=None, |
| model_name=None, |
| drift_check_baselines=None, |
| customer_metadata_properties=None, |
| domain=None, |
| sample_payload_url=None, |
| task=None, |
| framework=None, |
| framework_version=None, |
| nearest_model_name=None, |
| data_input_configuration=None, |
| **kwargs, |
| ): |
| """Creates a model package for creating SageMaker models or listing on Marketplace. |
| |
| Args: |
| content_types (list): The supported MIME types for the input data. |
| response_types (list): The supported MIME types for the output data. |
| inference_instances (list): A list of the instance types that are used to |
| generate inferences in real-time (default: None). |
| transform_instances (list): A list of the instance types on which a transformation |
| job can be run or on which an endpoint can be deployed (default: None). |
| image_uri (str): The container image uri for Model Package, if not specified, |
| Estimator's training container image will be used (default: None). |
| model_package_name (str): Model Package name, exclusive to `model_package_group_name`, |
| using `model_package_name` makes the Model Package un-versioned (default: None). |
| model_package_group_name (str): Model Package Group name, exclusive to |
| `model_package_name`, using `model_package_group_name` makes the Model Package |
| versioned (default: None). |
| model_metrics (ModelMetrics): ModelMetrics object (default: None). |
| metadata_properties (MetadataProperties): MetadataProperties (default: None). |
| marketplace_cert (bool): A boolean value indicating if the Model Package is certified |
| for AWS Marketplace (default: False). |
| approval_status (str): Model Approval Status, values can be "Approved", "Rejected", |
| or "PendingManualApproval" (default: "PendingManualApproval"). |
| description (str): Model Package description (default: None). |
| compile_model_family (str): Instance family for compiled model, if specified, a compiled |
| model will be used (default: None). |
| model_name (str): User defined model name (default: None). |
| drift_check_baselines (DriftCheckBaselines): DriftCheckBaselines object (default: None). |
| customer_metadata_properties (dict[str, str]): A dictionary of key-value paired |
| metadata properties (default: None). |
| domain (str): Domain values can be "COMPUTER_VISION", "NATURAL_LANGUAGE_PROCESSING", |
| "MACHINE_LEARNING" (default: None). |
| sample_payload_url (str): The S3 path where the sample payload is stored |
| (default: None). |
| task (str): Task values which are supported by Inference Recommender are "FILL_MASK", |
| "IMAGE_CLASSIFICATION", "OBJECT_DETECTION", "TEXT_GENERATION", "IMAGE_SEGMENTATION", |
| "CLASSIFICATION", "REGRESSION", "OTHER" (default: None). |
| framework (str): Machine learning framework of the model package container image |
| (default: None). |
| framework_version (str): Framework version of the Model Package Container Image |
| (default: None). |
| nearest_model_name (str): Name of a pre-trained machine learning benchmarked by |
| Amazon SageMaker Inference Recommender (default: None). |
| data_input_configuration (str): Input object for the model (default: None). |
| **kwargs: Passed to invocation of ``create_model()``. Implementations may customize |
| ``create_model()`` to accept ``**kwargs`` to customize model creation during |
| deploy. For more, see the implementation docs. |
| |
| Returns: |
| str: A string of SageMaker Model Package ARN. |
| """ |
| if isinstance(self.sagemaker_session, PipelineSession): |
| raise TypeError( |
| "estimator.register does not support PipelineSession at this moment. " |
| "Please use model.register with PipelineSession if you're using the ModelStep." |
| ) |
| default_name = name_from_base(self.base_job_name) |
| model_name = model_name or default_name |
| if compile_model_family is not None: |
| model = self._compiled_models[compile_model_family] |
| else: |
| if "model_kms_key" not in kwargs: |
| kwargs["model_kms_key"] = self.output_kms_key |
| model = self.create_model(image_uri=image_uri, **kwargs) |
| model.name = model_name |
| return model.register( |
| content_types, |
| response_types, |
| inference_instances, |
| transform_instances, |
| model_package_name, |
| model_package_group_name, |
| image_uri, |
| model_metrics, |
| metadata_properties, |
| marketplace_cert, |
| approval_status, |
| description, |
| drift_check_baselines=drift_check_baselines, |
| customer_metadata_properties=customer_metadata_properties, |
| domain=domain, |
| sample_payload_url=sample_payload_url, |
| task=task, |
| framework=framework, |
| framework_version=framework_version, |
| nearest_model_name=nearest_model_name, |
| data_input_configuration=data_input_configuration, |
| ) |
|
|
| @property |
| def model_data(self): |
| """str: The model location in S3. Only set if Estimator has been ``fit()``.""" |
| if self.latest_training_job is not None and not isinstance( |
| self.sagemaker_session, PipelineSession |
| ): |
| model_uri = self.sagemaker_session.sagemaker_client.describe_training_job( |
| TrainingJobName=self.latest_training_job.name |
| )["ModelArtifacts"]["S3ModelArtifacts"] |
| else: |
| logger.warning( |
| "No finished training job found associated with this estimator. Please make sure " |
| "this estimator is only used for building workflow config" |
| ) |
| model_uri = os.path.join( |
| self.output_path, self._current_job_name, "output", "model.tar.gz" |
| ) |
|
|
| return model_uri |
|
|
| @abstractmethod |
| def create_model(self, **kwargs): |
| """Create a SageMaker ``Model`` object that can be deployed to an ``Endpoint``. |
| |
| Args: |
| **kwargs: Keyword arguments used by the implemented method for |
| creating the ``Model``. |
| |
| Returns: |
| sagemaker.model.Model: A SageMaker ``Model`` object. See |
| :func:`~sagemaker.model.Model` for full details. |
| """ |
|
|
| @classmethod |
| def _prepare_init_params_from_job_description(cls, job_details, model_channel_name=None): |
| """Convert the job description to init params that can be handled by the class constructor. |
| |
| Args: |
| job_details: the returned job details from a describe_training_job |
| API call. |
| model_channel_name (str): Name of the channel where pre-trained |
| model data will be downloaded. |
| |
| Returns: |
| dictionary: The transformed init_params |
| """ |
| init_params = dict() |
|
|
| init_params["role"] = job_details["RoleArn"] |
| init_params["instance_count"] = job_details["ResourceConfig"]["InstanceCount"] |
| init_params["instance_type"] = job_details["ResourceConfig"]["InstanceType"] |
| init_params["volume_size"] = job_details["ResourceConfig"]["VolumeSizeInGB"] |
| init_params["max_run"] = job_details["StoppingCondition"]["MaxRuntimeInSeconds"] |
| init_params["input_mode"] = job_details["AlgorithmSpecification"]["TrainingInputMode"] |
| init_params["base_job_name"] = base_from_name(job_details["TrainingJobName"]) |
| init_params["output_path"] = job_details["OutputDataConfig"]["S3OutputPath"] |
| init_params["output_kms_key"] = job_details["OutputDataConfig"]["KmsKeyId"] |
| if "EnableNetworkIsolation" in job_details: |
| init_params["enable_network_isolation"] = job_details["EnableNetworkIsolation"] |
|
|
| if "KeepAlivePeriodInSeconds" in job_details["ResourceConfig"]: |
| init_params["keep_alive_period_in_seconds"] = job_details["ResourceConfig"][ |
| "keepAlivePeriodInSeconds" |
| ] |
|
|
| has_hps = "HyperParameters" in job_details |
| init_params["hyperparameters"] = job_details["HyperParameters"] if has_hps else {} |
|
|
| if "AlgorithmName" in job_details["AlgorithmSpecification"]: |
| init_params["algorithm_arn"] = job_details["AlgorithmSpecification"]["AlgorithmName"] |
| elif "TrainingImage" in job_details["AlgorithmSpecification"]: |
| init_params["image_uri"] = job_details["AlgorithmSpecification"]["TrainingImage"] |
| else: |
| raise RuntimeError( |
| "Invalid AlgorithmSpecification. Either TrainingImage or " |
| "AlgorithmName is expected. None was found." |
| ) |
|
|
| if "MetricDefinitons" in job_details["AlgorithmSpecification"]: |
| init_params["metric_definitions"] = job_details["AlgorithmSpecification"][ |
| "MetricsDefinition" |
| ] |
|
|
| if "EnableInterContainerTrafficEncryption" in job_details: |
| init_params["encrypt_inter_container_traffic"] = job_details[ |
| "EnableInterContainerTrafficEncryption" |
| ] |
|
|
| subnets, security_group_ids = vpc_utils.from_dict(job_details.get(vpc_utils.VPC_CONFIG_KEY)) |
| if subnets: |
| init_params["subnets"] = subnets |
| if security_group_ids: |
| init_params["security_group_ids"] = security_group_ids |
|
|
| if "InputDataConfig" in job_details and model_channel_name: |
| for channel in job_details["InputDataConfig"]: |
| if channel["ChannelName"] == model_channel_name: |
| init_params["model_channel_name"] = model_channel_name |
| init_params["model_uri"] = channel["DataSource"]["S3DataSource"]["S3Uri"] |
| break |
|
|
| if job_details.get("EnableManagedSpotTraining", False): |
| init_params["use_spot_instances"] = True |
| max_wait = job_details.get("StoppingCondition", {}).get("MaxWaitTimeInSeconds") |
| if max_wait: |
| init_params["max_wait"] = max_wait |
|
|
| if job_details.get("RetryStrategy", False): |
| init_params["max_retry_attempts"] = job_details.get("RetryStrategy", {}).get( |
| "MaximumRetryAttempts" |
| ) |
| max_wait = job_details.get("StoppingCondition", {}).get("MaxWaitTimeInSeconds") |
| if max_wait: |
| init_params["max_wait"] = max_wait |
| return init_params |
|
|
| def _get_instance_type(self): |
| """Determine the instance type to be used in the training_image_uri function. |
| |
| Returns: |
| instance_type: The instance_type to be used. |
| """ |
| if self.instance_type is not None: |
| return self.instance_type |
|
|
| if not isinstance(self.instance_groups, list) or len(self.instance_groups) == 0: |
| raise ValueError( |
| "instance_groups must be set if instance_type is not set and instance_groups " |
| "must be a list." |
| ) |
|
|
| for instance_group in self.instance_groups: |
| instance_type = instance_group.instance_type |
| if is_pipeline_variable(instance_type): |
| continue |
| match = re.match(r"^ml[\._]([a-z\d]+)\.?\w*$", instance_type) |
|
|
| if match: |
| family = match[1] |
| if family[0] in ("g", "p"): |
| return instance_type |
| else: |
| raise ValueError( |
| "Invalid SageMaker instance type for training with heterogeneous clusters: {}. " |
| "For options see: https://aws.amazon.com/sagemaker/pricing/instance-types".format( |
| instance_type |
| ) |
| ) |
|
|
| return self.instance_groups[0].instance_type |
|
|
| def transformer( |
| self, |
| instance_count, |
| instance_type, |
| strategy=None, |
| assemble_with=None, |
| output_path=None, |
| output_kms_key=None, |
| accept=None, |
| env=None, |
| max_concurrent_transforms=None, |
| max_payload=None, |
| tags=None, |
| role=None, |
| volume_kms_key=None, |
| vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT, |
| enable_network_isolation=None, |
| model_name=None, |
| ): |
| """Return a ``Transformer`` that uses a SageMaker Model based on the training job. |
| |
| It reuses the SageMaker Session and base job name used by |
| the Estimator. |
| |
| Args: |
| instance_count (int): Number of EC2 instances to use. |
| instance_type (str): Type of EC2 instance to use, for example, |
| 'ml.c4.xlarge'. |
| strategy (str): The strategy used to decide how to batch records in |
| a single request (default: None). Valid values: 'MultiRecord' |
| and 'SingleRecord'. |
| assemble_with (str): How the output is assembled (default: None). |
| Valid values: 'Line' or 'None'. |
| output_path (str): S3 location for saving the transform result. If |
| not specified, results are stored to a default bucket. |
| output_kms_key (str): Optional. KMS key ID for encrypting the |
| transform output (default: None). |
| accept (str): The accept header passed by the client to |
| the inference endpoint. If it is supported by the endpoint, |
| it will be the format of the batch transform output. |
| env (dict): Environment variables to be set for use during the |
| transform job (default: None). |
| max_concurrent_transforms (int): The maximum number of HTTP requests |
| to be made to each individual transform container at one time. |
| max_payload (int): Maximum size of the payload in a single HTTP |
| request to the container in MB. |
| tags (list[dict]): List of tags for labeling a transform job. If |
| none specified, then the tags used for the training job are used |
| for the transform job. |
| role (str): The ``ExecutionRoleArn`` IAM Role ARN for the ``Model``, |
| which is also used during transform jobs. If not specified, the |
| role from the Estimator will be used. |
| volume_kms_key (str): Optional. KMS key ID for encrypting the volume |
| attached to the ML compute instance (default: None). |
| vpc_config_override (dict[str, list[str]]): Optional override for the |
| VpcConfig set on the model. |
| Default: use subnets and security groups from this Estimator. |
| |
| * 'Subnets' (list[str]): List of subnet ids. |
| * 'SecurityGroupIds' (list[str]): List of security group ids. |
| |
| enable_network_isolation (bool): Specifies whether container will |
| run in network isolation mode. Network isolation mode restricts |
| the container access to outside networks (such as the internet). |
| The container does not make any inbound or outbound network |
| calls. If True, a channel named "code" will be created for any |
| user entry script for inference. Also known as Internet-free mode. |
| If not specified, this setting is taken from the estimator's |
| current configuration. |
| model_name (str): Name to use for creating an Amazon SageMaker |
| model. If not specified, the estimator generates a default job name |
| based on the training image name and current timestamp. |
| """ |
| tags = tags or self.tags |
| model_name = self._get_or_create_name(model_name) |
|
|
| if self.latest_training_job is None: |
| logger.warning( |
| "No finished training job found associated with this estimator. Please make sure " |
| "this estimator is only used for building workflow config" |
| ) |
| else: |
| if enable_network_isolation is None: |
| enable_network_isolation = self.enable_network_isolation() |
|
|
| model = self.create_model( |
| vpc_config_override=vpc_config_override, |
| model_kms_key=self.output_kms_key, |
| enable_network_isolation=enable_network_isolation, |
| ) |
|
|
| |
| model.name = model_name |
| if role is not None: |
| model.role = role |
|
|
| model._create_sagemaker_model(instance_type, tags=tags) |
|
|
| return Transformer( |
| model_name, |
| instance_count, |
| instance_type, |
| strategy=strategy, |
| assemble_with=assemble_with, |
| output_path=output_path, |
| output_kms_key=output_kms_key, |
| accept=accept, |
| max_concurrent_transforms=max_concurrent_transforms, |
| max_payload=max_payload, |
| env=env, |
| tags=tags, |
| base_transform_job_name=self.base_job_name, |
| volume_kms_key=volume_kms_key, |
| sagemaker_session=self.sagemaker_session, |
| ) |
|
|
| @property |
| def training_job_analytics(self): |
| """Return a ``TrainingJobAnalytics`` object for the current training job.""" |
| if self._current_job_name is None: |
| raise ValueError("Estimator is not associated with a TrainingJob") |
| return TrainingJobAnalytics( |
| self._current_job_name, sagemaker_session=self.sagemaker_session |
| ) |
|
|
| def get_vpc_config(self, vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT): |
| """Returns VpcConfig dict either from this Estimator's subnets and security groups. |
| |
| Or else validate and return an optional override value. |
| |
| Args: |
| vpc_config_override: |
| """ |
| if vpc_config_override is vpc_utils.VPC_CONFIG_DEFAULT: |
| return vpc_utils.to_dict(self.subnets, self.security_group_ids) |
| return vpc_utils.sanitize(vpc_config_override) |
|
|
| def _ensure_latest_training_job( |
| self, error_message="Estimator is not associated with a training job" |
| ): |
| """Placeholder docstring""" |
| if self.latest_training_job is None: |
| raise ValueError(error_message) |
|
|
| delete_endpoint = removed_function("delete_endpoint") |
|
|
| def enable_default_profiling(self): |
| """Update training job to enable Debugger monitoring. |
| |
| This method enables Debugger monitoring with |
| the default ``profiler_config`` parameter to collect system |
| metrics and the default built-in ``profiler_report`` rule. |
| Framework metrics won't be saved. |
| To update training job to emit framework metrics, you can use |
| :class:`~sagemaker.estimator.Estimator.update_profiler` |
| method and specify the framework metrics you want to enable. |
| |
| This method is callable when the training job is in progress while |
| Debugger monitoring is disabled. |
| """ |
| self._ensure_latest_training_job() |
| if not _region_supports_debugger(self.sagemaker_session.boto_region_name): |
| raise ValueError("Current region does not support profiler / debugger!") |
|
|
| training_job_details = self.latest_training_job.describe() |
|
|
| if training_job_details.get("ProfilingStatus") == "Enabled": |
| raise ValueError( |
| "Debugger monitoring is already enabled. To update the profiler_config parameter " |
| "and the Debugger profiling rules, please use the update_profiler function." |
| ) |
|
|
| if "ProfilerConfig" in training_job_details and training_job_details["ProfilerConfig"].get( |
| "S3OutputPath" |
| ): |
| self.profiler_config = ProfilerConfig( |
| s3_output_path=training_job_details["ProfilerConfig"]["S3OutputPath"] |
| ) |
| else: |
| self.profiler_config = ProfilerConfig(s3_output_path=self.output_path) |
|
|
| self.profiler_rules = [get_default_profiler_rule()] |
| self.profiler_rule_configs = self._prepare_profiler_rules() |
|
|
| _TrainingJob.update( |
| self, self.profiler_rule_configs, self.profiler_config._to_request_dict() |
| ) |
|
|
| def disable_profiling(self): |
| """Update the current training job in progress to disable profiling. |
| |
| Debugger stops collecting the system and framework metrics |
| and turns off the Debugger built-in monitoring and profiling rules. |
| |
| """ |
| self._ensure_latest_training_job() |
| if not _region_supports_debugger(self.sagemaker_session.boto_region_name): |
| raise ValueError("Current region does not support profiler / debugger!") |
|
|
| training_job_details = self.latest_training_job.describe() |
|
|
| if training_job_details.get("ProfilingStatus") == "Disabled": |
| raise ValueError("Profiler is already disabled.") |
|
|
| _TrainingJob.update( |
| self, profiler_config=ProfilerConfig._to_profiler_disabled_request_dict() |
| ) |
|
|
| def update_profiler( |
| self, |
| rules=None, |
| system_monitor_interval_millis=None, |
| s3_output_path=None, |
| framework_profile_params=None, |
| disable_framework_metrics=False, |
| ): |
| """Update training jobs to enable profiling. |
| |
| This method updates the ``profiler_config`` parameter |
| and initiates Debugger built-in rules for profiling. |
| |
| Args: |
| rules (list[:class:`~sagemaker.debugger.ProfilerRule`]): A list of |
| :class:`~sagemaker.debugger.ProfilerRule` objects to define |
| rules for continuous analysis with SageMaker Debugger. Currently, you can |
| only add new profiler rules during the training job. (default: None) |
| s3_output_path (str): The location in S3 to store the output. If profiler is enabled |
| once, s3_output_path cannot be changed. (default: None) |
| system_monitor_interval_millis (int): How often profiling system metrics are |
| collected; Unit: Milliseconds (default: None) |
| framework_profile_params (:class:`~sagemaker.debugger.FrameworkProfile`): |
| A parameter object for framework metrics profiling. Configure it using |
| the :class:`~sagemaker.debugger.FrameworkProfile` class. |
| To use the default framework profile parameters, pass ``FrameworkProfile()``. |
| For more information about the default values, |
| see :class:`~sagemaker.debugger.FrameworkProfile`. (default: None) |
| disable_framework_metrics (bool): Specify whether to disable all the framework metrics. |
| This won't update system metrics and the Debugger built-in rules for monitoring. |
| To stop both monitoring and profiling, |
| use the :class:`~sagemaker.estimator.Estimator.desable_profiling` |
| method. (default: ``False``) |
| |
| .. attention:: |
| |
| Updating the profiling configuration for TensorFlow dataloader profiling |
| is currently not available. If you started a TensorFlow training job only with |
| monitoring and want to enable profiling while the training job is running, |
| the dataloader profiling cannot be updated. |
| |
| """ |
| self._ensure_latest_training_job() |
| if not _region_supports_debugger(self.sagemaker_session.boto_region_name): |
| raise ValueError("Current region does not support profiler / debugger!") |
|
|
| if ( |
| not rules |
| and not system_monitor_interval_millis |
| and not s3_output_path |
| and not framework_profile_params |
| and not disable_framework_metrics |
| ): |
| raise ValueError("Please provide profiler config or profiler rule to be updated.") |
|
|
| if disable_framework_metrics and framework_profile_params: |
| raise ValueError( |
| "framework_profile_params cannot be set when disable_framework_metrics is True" |
| ) |
|
|
| profiler_config_request_dict = None |
| profiler_rule_configs = None |
|
|
| if rules: |
| for rule in rules: |
| if not isinstance(rule, ProfilerRule): |
| raise ValueError("Please provide ProfilerRule to be updated.") |
| self.profiler_rules = rules |
| profiler_rule_configs = self._prepare_profiler_rules() |
|
|
| if disable_framework_metrics: |
| empty_framework_profile_param = FrameworkProfile() |
| empty_framework_profile_param.profiling_parameters = {} |
| self.profiler_config = ProfilerConfig( |
| s3_output_path=s3_output_path, |
| system_monitor_interval_millis=system_monitor_interval_millis, |
| framework_profile_params=empty_framework_profile_param, |
| ) |
| else: |
| self.profiler_config = ProfilerConfig( |
| s3_output_path=s3_output_path, |
| system_monitor_interval_millis=system_monitor_interval_millis, |
| framework_profile_params=framework_profile_params, |
| ) |
|
|
| profiler_config_request_dict = self.profiler_config._to_request_dict() |
|
|
| _TrainingJob.update(self, profiler_rule_configs, profiler_config_request_dict) |
|
|
|
|
| class _TrainingJob(_Job): |
| """Placeholder docstring""" |
|
|
| @classmethod |
| def start_new(cls, estimator, inputs, experiment_config): |
| """Create a new Amazon SageMaker training job from the estimator. |
| |
| Args: |
| estimator (sagemaker.estimator.EstimatorBase): Estimator object |
| created by the user. |
| inputs (str): Parameters used when called |
| :meth:`~sagemaker.estimator.EstimatorBase.fit`. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| Returns: |
| sagemaker.estimator._TrainingJob: Constructed object that captures |
| all information about the started training job. |
| """ |
| train_args = cls._get_train_args(estimator, inputs, experiment_config) |
|
|
| estimator.sagemaker_session.train(**train_args) |
|
|
| return cls(estimator.sagemaker_session, estimator._current_job_name) |
|
|
| @classmethod |
| def _get_train_args(cls, estimator, inputs, experiment_config): |
| """Constructs a dict of arguments for an Amazon SageMaker training job from the estimator. |
| |
| Args: |
| estimator (sagemaker.estimator.EstimatorBase): Estimator object |
| created by the user. |
| inputs (str): Parameters used when called |
| :meth:`~sagemaker.estimator.EstimatorBase.fit`. |
| experiment_config (dict[str, str]): Experiment management configuration. |
| Optionally, the dict can contain three keys: |
| 'ExperimentName', 'TrialName', and 'TrialComponentDisplayName'. |
| The behavior of setting these keys is as follows: |
| * If `ExperimentName` is supplied but `TrialName` is not a Trial will be |
| automatically created and the job's Trial Component associated with the Trial. |
| * If `TrialName` is supplied and the Trial already exists the job's Trial Component |
| will be associated with the Trial. |
| * If both `ExperimentName` and `TrialName` are not supplied the trial component |
| will be unassociated. |
| * `TrialComponentDisplayName` is used for display in Studio. |
| |
| Returns: |
| Dict: dict for `sagemaker.session.Session.train` method |
| """ |
|
|
| local_mode = estimator.sagemaker_session.local_mode |
| model_uri = estimator.model_uri |
|
|
| |
| if cls._is_local_channel(inputs) or cls._is_local_channel(model_uri): |
| if not local_mode: |
| raise ValueError( |
| "File URIs are supported in local mode only. Please use a S3 URI instead." |
| ) |
|
|
| config = _Job._load_config(inputs, estimator) |
|
|
| current_hyperparameters = estimator.hyperparameters() |
| if current_hyperparameters is not None: |
| hyperparameters = {str(k): to_string(v) for (k, v) in current_hyperparameters.items()} |
|
|
| train_args = config.copy() |
| train_args["input_mode"] = estimator.input_mode |
| train_args["job_name"] = estimator._current_job_name |
| train_args["hyperparameters"] = hyperparameters |
| train_args["tags"] = estimator.tags |
| train_args["metric_definitions"] = estimator.metric_definitions |
| train_args["experiment_config"] = experiment_config |
| train_args["environment"] = estimator.environment |
|
|
| if isinstance(inputs, TrainingInput): |
| if "InputMode" in inputs.config: |
| logger.debug( |
| "Selecting TrainingInput's input_mode (%s) for TrainingInputMode.", |
| inputs.config["InputMode"], |
| ) |
| train_args["input_mode"] = inputs.config["InputMode"] |
|
|
| |
| |
| if estimator.enable_network_isolation(): |
| train_args["enable_network_isolation"] = estimator.enable_network_isolation() |
|
|
| if estimator.max_retry_attempts is not None: |
| train_args["retry_strategy"] = {"MaximumRetryAttempts": estimator.max_retry_attempts} |
| else: |
| train_args["retry_strategy"] = None |
|
|
| |
| |
| if estimator.encrypt_inter_container_traffic: |
| train_args[ |
| "encrypt_inter_container_traffic" |
| ] = estimator.encrypt_inter_container_traffic |
|
|
| if isinstance(estimator, sagemaker.algorithm.AlgorithmEstimator): |
| train_args["algorithm_arn"] = estimator.algorithm_arn |
| else: |
| train_args["image_uri"] = estimator.training_image_uri() |
|
|
| if estimator.debugger_rule_configs: |
| train_args["debugger_rule_configs"] = estimator.debugger_rule_configs |
|
|
| if estimator.debugger_hook_config: |
| estimator.debugger_hook_config.collection_configs = estimator.collection_configs |
| train_args["debugger_hook_config"] = estimator.debugger_hook_config._to_request_dict() |
|
|
| if estimator.tensorboard_output_config: |
| train_args[ |
| "tensorboard_output_config" |
| ] = estimator.tensorboard_output_config._to_request_dict() |
|
|
| cls._add_spot_checkpoint_args(local_mode, estimator, train_args) |
|
|
| if estimator.enable_sagemaker_metrics is not None: |
| train_args["enable_sagemaker_metrics"] = estimator.enable_sagemaker_metrics |
|
|
| if estimator.profiler_rule_configs: |
| train_args["profiler_rule_configs"] = estimator.profiler_rule_configs |
|
|
| if estimator.profiler_config: |
| train_args["profiler_config"] = estimator.profiler_config._to_request_dict() |
|
|
| return train_args |
|
|
| @classmethod |
| def _add_spot_checkpoint_args(cls, local_mode, estimator, train_args): |
| """Placeholder docstring""" |
| if estimator.use_spot_instances: |
| if local_mode: |
| raise ValueError("Spot training is not supported in local mode.") |
| |
| |
| train_args["use_spot_instances"] = estimator.use_spot_instances |
|
|
| if estimator.checkpoint_s3_uri: |
| if local_mode: |
| raise ValueError("Setting checkpoint_s3_uri is not supported in local mode.") |
| train_args["checkpoint_s3_uri"] = estimator.checkpoint_s3_uri |
|
|
| if estimator.checkpoint_local_path: |
| if local_mode: |
| raise ValueError("Setting checkpoint_local_path is not supported in local mode.") |
| train_args["checkpoint_local_path"] = estimator.checkpoint_local_path |
|
|
| @classmethod |
| def _is_local_channel(cls, input_uri): |
| """Placeholder docstring""" |
| return isinstance(input_uri, string_types) and input_uri.startswith("file://") |
|
|
| @classmethod |
| def update( |
| cls, estimator, profiler_rule_configs=None, profiler_config=None, resource_config=None |
| ): |
| """Update a running Amazon SageMaker training job. |
| |
| Args: |
| estimator (sagemaker.estimator.EstimatorBase): Estimator object created by the user. |
| profiler_rule_configs (list): List of profiler rule configurations to be |
| updated in the training job. (default: None). |
| profiler_config (dict): Configuration for how profiling information is emitted with |
| SageMaker Debugger. (default: None). |
| resource_config (dict): Configuration of the resources for the training job. You can |
| update the keep-alive period if the warm pool status is `Available`. No other fields |
| can be updated. (default: None). |
| |
| Returns: |
| sagemaker.estimator._TrainingJob: Constructed object that captures |
| all information about the updated training job. |
| """ |
| update_args = cls._get_update_args( |
| estimator, profiler_rule_configs, profiler_config, resource_config |
| ) |
| estimator.sagemaker_session.update_training_job(**update_args) |
|
|
| return estimator.latest_training_job |
|
|
| @classmethod |
| def _get_update_args(cls, estimator, profiler_rule_configs, profiler_config, resource_config): |
| """Constructs a dict of arguments for updating an Amazon SageMaker training job. |
| |
| Args: |
| estimator (sagemaker.estimator.EstimatorBase): Estimator object |
| created by the user. |
| profiler_rule_configs (list): List of profiler rule configurations to be |
| updated in the training job. (default: None). |
| profiler_config (dict): Configuration for how profiling information is emitted with |
| SageMaker Debugger. (default: None). |
| resource_config (dict): Configuration of the resources for the training job. You can |
| update the keep-alive period if the warm pool status is `Available`. No other fields |
| can be updated. (default: None). |
| |
| Returns: |
| Dict: dict for `sagemaker.session.Session.update_training_job` method |
| """ |
| update_args = {"job_name": estimator.latest_training_job.name} |
| update_args.update(build_dict("profiler_rule_configs", profiler_rule_configs)) |
| update_args.update(build_dict("profiler_config", profiler_config)) |
| update_args.update(build_dict("resource_config", resource_config)) |
|
|
| return update_args |
|
|
| def wait(self, logs="All"): |
| """Placeholder docstring. |
| |
| Args: |
| logs ([str]): A list of strings specifying which logs to print. Acceptable |
| strings are "All", "None", "Training", or "Rules". To maintain backwards |
| compatibility, boolean values are also accepted and converted to strings. |
| """ |
| |
| log_string_map = {True: "All", False: "None"} |
| if isinstance(logs, bool): |
| logs = log_string_map[logs] |
| |
| if logs != "None": |
| self.sagemaker_session.logs_for_job(self.job_name, wait=True, log_type=logs) |
| else: |
| self.sagemaker_session.wait_for_job(self.job_name) |
|
|
| def describe(self): |
| """Returns a response from the DescribeTrainingJob API call.""" |
| return self.sagemaker_session.describe_training_job(self.job_name) |
|
|
| def rule_job_summary(self): |
| """Calls describe_training_job and returns two dictionaries. |
| |
| Returns: |
| list[dict]: A list of DebugRuleEvaluationStatuses and ProfilerRuleEvaluationStatuses |
| dictionary. |
| """ |
| job_summary = self.describe() |
| rule_eval_statuses = job_summary.get("DebugRuleEvaluationStatuses") or [] |
| rule_eval_statuses.extend(job_summary.get("ProfilerRuleEvaluationStatuses") or []) |
|
|
| return rule_eval_statuses |
|
|
| def stop(self): |
| """Stops the training job.""" |
| self.sagemaker_session.stop_training_job(self.name) |
|
|
|
|
| class Estimator(EstimatorBase): |
| """A generic Estimator to train using any supplied algorithm. |
| |
| This class is designed for use with algorithms that don't have their own, custom class. |
| """ |
|
|
| def __init__( |
| self, |
| image_uri: Union[str, PipelineVariable], |
| role: str, |
| instance_count: Optional[Union[int, PipelineVariable]] = None, |
| instance_type: Optional[Union[str, PipelineVariable]] = None, |
| keep_alive_period_in_seconds: Optional[Union[int, PipelineVariable]] = None, |
| volume_size: Union[int, PipelineVariable] = 30, |
| volume_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| max_run: Union[int, PipelineVariable] = 24 * 60 * 60, |
| input_mode: Union[str, PipelineVariable] = "File", |
| output_path: Optional[Union[str, PipelineVariable]] = None, |
| output_kms_key: Optional[Union[str, PipelineVariable]] = None, |
| base_job_name: Optional[str] = None, |
| sagemaker_session: Optional[Session] = None, |
| hyperparameters: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| tags: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| subnets: Optional[List[Union[str, PipelineVariable]]] = None, |
| security_group_ids: Optional[List[Union[str, PipelineVariable]]] = None, |
| model_uri: Optional[str] = None, |
| model_channel_name: Union[str, PipelineVariable] = "model", |
| metric_definitions: Optional[List[Dict[str, Union[str, PipelineVariable]]]] = None, |
| encrypt_inter_container_traffic: Union[bool, PipelineVariable] = False, |
| use_spot_instances: Union[bool, PipelineVariable] = False, |
| max_wait: Optional[Union[int, PipelineVariable]] = None, |
| checkpoint_s3_uri: Optional[Union[str, PipelineVariable]] = None, |
| checkpoint_local_path: Optional[Union[str, PipelineVariable]] = None, |
| enable_network_isolation: Union[bool, PipelineVariable] = False, |
| rules: Optional[List[RuleBase]] = None, |
| debugger_hook_config: Optional[Union[DebuggerHookConfig, bool]] = None, |
| tensorboard_output_config: Optional[TensorBoardOutputConfig] = None, |
| enable_sagemaker_metrics: Optional[Union[bool, PipelineVariable]] = None, |
| profiler_config: Optional[ProfilerConfig] = None, |
| disable_profiler: bool = False, |
| environment: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| max_retry_attempts: Optional[Union[int, PipelineVariable]] = None, |
| source_dir: Optional[Union[str, PipelineVariable]] = None, |
| git_config: Optional[Dict[str, str]] = None, |
| container_log_level: Union[int, PipelineVariable] = logging.INFO, |
| code_location: Optional[str] = None, |
| entry_point: Optional[Union[str, PipelineVariable]] = None, |
| dependencies: Optional[List[str]] = None, |
| instance_groups: Optional[List[InstanceGroup]] = None, |
| **kwargs, |
| ): |
| """Initialize an ``Estimator`` instance. |
| |
| Args: |
| image_uri (str or PipelineVariable): The container image to use for training. |
| role (str): An AWS IAM role (either name or full ARN). The Amazon |
| SageMaker training jobs and APIs that create Amazon SageMaker |
| endpoints use this role to access training data and model |
| artifacts. After the endpoint is created, the inference code |
| might use the IAM role, if it needs to access an AWS resource. |
| instance_count (int or PipelineVariable): Number of Amazon EC2 instances to use |
| for training. Required if instance_groups is not set. |
| instance_type (str or PipelineVariable): Type of EC2 instance to use for training, |
| for example, ``'ml.c4.xlarge'``. Required if instance_groups is |
| not set. |
| keep_alive_period_in_seconds (int): The duration of time in seconds |
| to retain configured resources in a warm pool for subsequent |
| training jobs (default: None). |
| volume_size (int or PipelineVariable): Size in GB of the storage volume to use for |
| storing input and output data during training (default: 30). |
| |
| Must be large enough to store training data if File mode is |
| used, which is the default mode. |
| |
| When you use an ML instance with the EBS-only storage option |
| such as ``ml.c5`` and ``ml.p2``, |
| you must define the size of the EBS |
| volume through the ``volume_size`` parameter in the estimator class. |
| |
| .. note:: |
| |
| When you use an ML instance with `NVMe SSD volumes |
| <https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ssd-instance-store.html#nvme-ssd-volumes>`_ |
| such as ``ml.p4d``, ``ml.g4dn``, and ``ml.g5``, |
| do not include this parameter in the estimator configuration. |
| If you use one of those ML instance types, |
| SageMaker doesn't provision Amazon EBS General Purpose SSD |
| (gp2) storage nor take this parameter to adjust the NVMe instance storage. |
| Available storage is fixed to the NVMe instance storage |
| capacity. SageMaker configures storage paths for training |
| datasets, checkpoints, model artifacts, and outputs to use the |
| entire capacity of the instance storage. |
| |
| Note that if you include this parameter and specify a number that |
| exceeds the size of the NVMe volume attached to the instance type, |
| SageMaker returns an ``Invalid VolumeSizeInGB`` error. |
| |
| To look up instance types and their instance storage types |
| and volumes, see `Amazon EC2 Instance Types |
| <http://aws.amazon.com/ec2/instance-types/>`_. |
| |
| To find the default local paths defined by the SageMaker |
| training platform, see `Amazon SageMaker Training Storage |
| Folders for Training Datasets, Checkpoints, Model Artifacts, |
| and Outputs |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/model-train-storage.html>`_. |
| volume_kms_key (str or PipelineVariable): Optional. KMS key ID for encrypting EBS |
| volume attached to the training instance (default: None). |
| max_run (int or PipelineVariable): Timeout in seconds for training (default: 24 * |
| 60 * 60). After this amount of time Amazon SageMaker terminates |
| the job regardless of its current status. |
| input_mode (str or PipelineVariable): The input mode that the algorithm supports |
| (default: 'File'). Valid modes: |
| |
| * 'File' - Amazon SageMaker copies the training dataset from the |
| S3 location to a local directory. |
| * 'Pipe' - Amazon SageMaker streams data directly from S3 to the |
| container via a Unix-named pipe. |
| |
| This argument can be overriden on a per-channel basis using |
| ``sagemaker.inputs.TrainingInput.input_mode``. |
| output_path (str or PipelineVariable): S3 location for saving the training result |
| (model artifacts and output files). If not specified, results are |
| stored to a default bucket. If the bucket with the specific name |
| does not exist, the estimator creates the bucket during the |
| :meth:`~sagemaker.estimator.EstimatorBase.fit` method execution. |
| output_kms_key (str or PipelineVariable): Optional. KMS key ID for encrypting the |
| training output (default: None). |
| base_job_name (str): Prefix for training job name when the |
| :meth:`~sagemaker.estimator.EstimatorBase.fit` method launches. |
| If not specified, the estimator generates a default job name, |
| based on the training image name and current timestamp. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the estimator creates one |
| using the default AWS configuration chain. |
| hyperparameters (dict[str, str] or dict[str, PipelineVariable]): |
| Dictionary containing the hyperparameters to initialize this estimator with. |
| tags (list[dict[str, str] or list[dict[str, PipelineVariable]]): List of tags for |
| labeling a training job. For more, see |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_Tag.html. |
| subnets (list[str] or list[PipelineVariable]): List of subnet ids. |
| If not specified training job will be created without VPC config. |
| security_group_ids (list[str] or list[PipelineVariable]): List of security group ids. |
| If not specified training job will be created without VPC config. |
| model_uri (str): URI where a pre-trained model is stored, either |
| locally or in S3 (default: None). If specified, the estimator |
| will create a channel pointing to the model so the training job |
| can download it. This model can be a 'model.tar.gz' from a |
| previous training job, or other artifacts coming from a |
| different source. |
| |
| In local mode, this should point to the path in which the model |
| is located and not the file itself, as local Docker containers |
| will try to mount the URI as a volume. |
| |
| More information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/cdf-training.html#td-deserialization |
| model_channel_name (str or PipelineVariable): Name of the channel where 'model_uri' will |
| be downloaded (default: 'model'). |
| metric_definitions (list[dict[str, str] or list[dict[str, PipelineVariable]]): A list of |
| dictionaries that defines the metric(s) used to evaluate the training jobs. Each |
| dictionary contains two keys: 'Name' for the name of the metric, |
| and 'Regex' for the regular expression used to extract the |
| metric from the logs. This should be defined only for jobs that |
| don't use an Amazon algorithm. |
| encrypt_inter_container_traffic (bool or PipelineVariable): Specifies whether traffic |
| between training containers is encrypted for the training job |
| (default: ``False``). |
| use_spot_instances (bool or PipelineVariable): Specifies whether to use SageMaker |
| Managed Spot instances for training. If enabled then the |
| ``max_wait`` arg should also be set. |
| |
| More information: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/model-managed-spot-training.html |
| (default: ``False``). |
| max_wait (int or PipelineVariable): Timeout in seconds waiting for spot training |
| job (default: None). After this amount of time Amazon |
| SageMaker will stop waiting for managed spot training job to |
| complete (default: None). |
| checkpoint_s3_uri (str or PipelineVariable): The S3 URI in which to persist checkpoints |
| that the algorithm persists (if any) during training. (default: |
| None). |
| checkpoint_local_path (str or PipelineVariable): The local path that the algorithm |
| writes its checkpoints to. SageMaker will persist all files |
| under this path to `checkpoint_s3_uri` continually during |
| training. On job startup the reverse happens - data from the |
| s3 location is downloaded to this path before the algorithm is |
| started. If the path is unset then SageMaker assumes the |
| checkpoints will be provided under `/opt/ml/checkpoints/`. |
| (default: None). |
| enable_network_isolation (bool or PipelineVariable): Specifies whether container will |
| run in network isolation mode (default: ``False``). Network |
| isolation mode restricts the container access to outside networks |
| (such as the Internet). The container does not make any inbound or |
| outbound network calls. Also known as Internet-free mode. |
| rules (list[:class:`~sagemaker.debugger.RuleBase`]): A list of |
| :class:`~sagemaker.debugger.RuleBase` objects used to define |
| SageMaker Debugger rules for real-time analysis |
| (default: None). For more information, |
| see `Continuous analyses through rules |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html |
| #continuous-analyses-through-rules)>`_. |
| debugger_hook_config (:class:`~sagemaker.debugger.DebuggerHookConfig` or bool): |
| Configuration for how debugging information is emitted with |
| SageMaker Debugger. If not specified, a default one is created using |
| the estimator's ``output_path``, unless the region does not |
| support SageMaker Debugger. To disable SageMaker Debugger, |
| set this parameter to ``False``. For more information, see |
| `Capture real-time debugging data during model training in Amazon SageMaker |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html# |
| capture-real-time-debugging-data-during-model-training-in-amazon-sagemaker>`_. |
| tensorboard_output_config (:class:`~sagemaker.debugger.TensorBoardOutputConfig`): |
| Configuration for customizing debugging visualization using TensorBoard |
| (default: None). For more information, |
| see `Capture real time tensorboard data |
| <https://sagemaker.readthedocs.io/en/stable/amazon_sagemaker_debugger.html# |
| capture-real-time-tensorboard-data-from-the-debugging-hook>`_. |
| enable_sagemaker_metrics (bool or PipelineVariable): enable SageMaker Metrics Time |
| Series. For more information, see `AlgorithmSpecification API |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/ |
| API_AlgorithmSpecification.html#SageMaker-Type-AlgorithmSpecification- |
| EnableSageMakerMetricsTimeSeries>`_. |
| (default: None). |
| profiler_config (:class:`~sagemaker.debugger.ProfilerConfig`): |
| Configuration for how SageMaker Debugger collects |
| monitoring and profiling information from your training job. |
| If not specified, Debugger will be configured with |
| a default configuration and will save system and framework metrics |
| the estimator's default ``output_path`` in Amazon S3. |
| Use :class:`~sagemaker.debugger.ProfilerConfig` to configure this parameter. |
| To disable SageMaker Debugger monitoring and profiling, set the |
| ``disable_profiler`` parameter to ``True``. |
| disable_profiler (bool): Specifies whether Debugger monitoring and profiling |
| will be disabled (default: ``False``). |
| environment (dict[str, str] or dict[str, PipelineVariable]) : Environment variables to |
| be set for use during training job (default: None) |
| max_retry_attempts (int or PipelineVariable): The number of times to move a job |
| to the STARTING status. You can specify between 1 and 30 attempts. |
| If the value of attempts is greater than zero, |
| the job is retried on InternalServerFailure |
| the same number of attempts as the value. |
| You can cap the total duration for your job by setting ``max_wait`` and ``max_run`` |
| (default: None) |
| source_dir (str or PipelineVariable): The absolute, relative, or S3 URI Path to |
| a directory with any other training source code dependencies aside from the entry |
| point file (default: None). If ``source_dir`` is an S3 URI, it must |
| point to a tar.gz file. Structure within this directory is preserved |
| when training on Amazon SageMaker. If 'git_config' is provided, |
| 'source_dir' should be a relative location to a directory in the Git |
| repo. |
| |
| .. admonition:: Example |
| |
| With the following GitHub repo directory structure: |
| |
| >>> |----- README.md |
| >>> |----- src |
| >>> |----- train.py |
| >>> |----- test.py |
| |
| if you need 'train.py' |
| as the entry point and 'test.py' as the training source code, you can assign |
| entry_point='train.py', source_dir='src'. |
| git_config (dict[str, str]): Git configurations used for cloning |
| files, including ``repo``, ``branch``, ``commit``, |
| ``2FA_enabled``, ``username``, ``password`` and ``token``. The |
| ``repo`` field is required. All other fields are optional. |
| ``repo`` specifies the Git repository where your training script |
| is stored. If you don't provide ``branch``, the default value |
| 'master' is used. If you don't provide ``commit``, the latest |
| commit in the specified branch is used. |
| |
| .. admonition:: Example |
| |
| The following config: |
| |
| >>> git_config = {'repo': 'https://github.com/aws/sagemaker-python-sdk.git', |
| >>> 'branch': 'test-branch-git-config', |
| >>> 'commit': '329bfcf884482002c05ff7f44f62599ebc9f445a'} |
| |
| results in cloning the repo specified in 'repo', then |
| checking out the 'master' branch, and checking out the specified |
| commit. |
| |
| ``2FA_enabled``, ``username``, ``password`` and ``token`` are |
| used for authentication. For GitHub (or other Git) accounts, set |
| ``2FA_enabled`` to 'True' if two-factor authentication is |
| enabled for the account, otherwise set it to 'False'. If you do |
| not provide a value for ``2FA_enabled``, a default value of |
| 'False' is used. CodeCommit does not support two-factor |
| authentication, so do not provide "2FA_enabled" with CodeCommit |
| repositories. |
| |
| For GitHub and other Git repos, when SSH URLs are provided, it |
| doesn't matter whether 2FA is enabled or disabled. You should |
| either have no passphrase for the SSH key pairs or have the |
| ssh-agent configured so that you will not be prompted for the SSH |
| passphrase when you run the 'git clone' command with SSH URLs. When |
| HTTPS URLs are provided, if 2FA is disabled, then either ``token`` |
| or ``username`` and ``password`` are be used for authentication if provided. |
| ``Token`` is prioritized. If 2FA is enabled, only ``token`` is used |
| for authentication if provided. If required authentication info |
| is not provided, the SageMaker Python SDK attempts to use local credentials |
| to authenticate. If that fails, an error message is thrown. |
| |
| For CodeCommit repos, 2FA is not supported, so ``2FA_enabled`` |
| should not be provided. There is no token in CodeCommit, so |
| ``token`` should also not be provided. When ``repo`` is an SSH URL, |
| the requirements are the same as GitHub repos. When ``repo`` |
| is an HTTPS URL, ``username`` and ``password`` are used for |
| authentication if they are provided. If they are not provided, |
| the SageMaker Python SDK attempts to use either the CodeCommit |
| credential helper or local credential storage for authentication. |
| container_log_level (int or PipelineVariable): The log level to use within the container |
| (default: logging.INFO). Valid values are defined in the Python |
| logging module. |
| code_location (str): The S3 prefix URI where custom code is |
| uploaded (default: None). You must not include a trailing slash because |
| a string prepended with a "/" is appended to ``code_location``. The code |
| file uploaded to S3 is 'code_location/job-name/source/sourcedir.tar.gz'. |
| If not specified, the default ``code location`` is 's3://output_bucket/job-name/'. |
| entry_point (str or PipelineVariable): The absolute or relative path to the local Python |
| source file that should be executed as the entry point to |
| training. If ``source_dir`` is specified, then ``entry_point`` |
| must point to a file located at the root of ``source_dir``. |
| If 'git_config' is provided, 'entry_point' should be |
| a relative location to the Python source file in the Git repo. |
| |
| Example: |
| With the following GitHub repo directory structure: |
| |
| >>> |----- README.md |
| >>> |----- src |
| >>> |----- train.py |
| >>> |----- test.py |
| |
| You can assign entry_point='src/train.py'. |
| dependencies (list[str]): A list of absolute or relative paths to directories |
| with any additional libraries that should be exported |
| to the container (default: []). The library folders are |
| copied to SageMaker in the same folder where the entrypoint is |
| copied. If 'git_config' is provided, 'dependencies' should be a |
| list of relative locations to directories with any additional |
| libraries needed in the Git repo. |
| |
| .. admonition:: Example |
| |
| The following Estimator call: |
| |
| >>> Estimator(entry_point='train.py', |
| ... dependencies=['my/libs/common', 'virtual-env']) |
| |
| results in the following structure inside the container: |
| |
| >>> $ ls |
| |
| >>> opt/ml/code |
| >>> |------ train.py |
| >>> |------ common |
| >>> |------ virtual-env |
| |
| This is not supported with "local code" in Local Mode. |
| instance_groups (list[:class:`sagemaker.instance_group.InstanceGroup`]): |
| Optional. A list of ``InstanceGroup`` objects |
| for launching a training job with a heterogeneous cluster. |
| For example: |
| |
| .. code:: python |
| |
| instance_groups=[ |
| sagemaker.InstanceGroup( |
| 'instance_group_name_1', 'ml.p3dn.24xlarge', 64), |
| sagemaker.InstanceGroup( |
| 'instance_group_name_2', 'ml.c5n.18xlarge', 64)] |
| |
| For instructions on how to use ``InstanceGroup`` objects |
| to configure a heterogeneous cluster |
| through the SageMaker generic and framework estimator classes, see |
| `Train Using a Heterogeneous Cluster |
| <https://docs.aws.amazon.com/sagemaker/latest/dg/train-heterogeneous-cluster.html>`_ |
| in the *Amazon SageMaker developer guide*. |
| """ |
| self.image_uri = image_uri |
| self._hyperparameters = hyperparameters.copy() if hyperparameters else {} |
| super(Estimator, self).__init__( |
| role, |
| instance_count, |
| instance_type, |
| keep_alive_period_in_seconds, |
| volume_size, |
| volume_kms_key, |
| max_run, |
| input_mode, |
| output_path, |
| output_kms_key, |
| base_job_name, |
| sagemaker_session, |
| tags, |
| subnets, |
| security_group_ids, |
| model_uri=model_uri, |
| model_channel_name=model_channel_name, |
| metric_definitions=metric_definitions, |
| encrypt_inter_container_traffic=encrypt_inter_container_traffic, |
| use_spot_instances=use_spot_instances, |
| max_wait=max_wait, |
| checkpoint_s3_uri=checkpoint_s3_uri, |
| checkpoint_local_path=checkpoint_local_path, |
| rules=rules, |
| debugger_hook_config=debugger_hook_config, |
| tensorboard_output_config=tensorboard_output_config, |
| enable_sagemaker_metrics=enable_sagemaker_metrics, |
| enable_network_isolation=enable_network_isolation, |
| profiler_config=profiler_config, |
| disable_profiler=disable_profiler, |
| environment=environment, |
| max_retry_attempts=max_retry_attempts, |
| container_log_level=container_log_level, |
| source_dir=source_dir, |
| git_config=git_config, |
| code_location=code_location, |
| entry_point=entry_point, |
| dependencies=dependencies, |
| hyperparameters=hyperparameters, |
| instance_groups=instance_groups, |
| **kwargs, |
| ) |
|
|
| self.set_hyperparameters(**self._hyperparameters) |
|
|
| def training_image_uri(self): |
| """Returns the docker image to use for training. |
| |
| The fit() method, that does the model training, calls this method to |
| find the image to use for model training. |
| """ |
| return self.image_uri |
|
|
| def set_hyperparameters(self, **kwargs): |
| """Sets the hyperparameter dictionary to use for training. |
| |
| The hyperparameters are made accessible as a dict[str, str] to the |
| training code on SageMaker. For convenience, this accepts other types |
| for keys and values, but ``str()`` will be called to convert them before |
| training. |
| |
| If a source directory is specified, this method escapes the dict argument as JSON, |
| and updates the private hyperparameter attribute. |
| """ |
| if self.source_dir: |
| self._hyperparameters.update(EstimatorBase._json_encode_hyperparameters(kwargs)) |
| else: |
| for k, v in kwargs.items(): |
| self._hyperparameters[k] = v |
|
|
| def hyperparameters(self): |
| """Returns the hyperparameters as a dictionary to use for training. |
| |
| The fit() method, that does the model training, calls this method to |
| find the hyperparameters you specified. |
| """ |
| return self._hyperparameters |
|
|
| def create_model( |
| self, |
| role=None, |
| image_uri=None, |
| predictor_cls=None, |
| vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT, |
| **kwargs, |
| ): |
| """Create a model to deploy. |
| |
| The serializer and deserializer arguments are only used to define a |
| default Predictor. They are ignored if an explicit predictor class is passed in. |
| Other arguments are passed through to the Model class. |
| |
| Args: |
| role (str): The ``ExecutionRoleArn`` IAM Role ARN for the ``Model``, |
| which is also used during transform jobs. If not specified, the |
| role from the Estimator will be used. |
| image_uri (str): A Docker image URI to use for deploying the model. |
| Defaults to the image used for training. |
| predictor_cls (Predictor): The predictor class to use when |
| deploying the model. |
| vpc_config_override (dict[str, list[str]]): Optional override for VpcConfig set on |
| the model. |
| Default: use subnets and security groups from this Estimator. |
| * 'Subnets' (list[str]): List of subnet ids. |
| * 'SecurityGroupIds' (list[str]): List of security group ids. |
| **kwargs: Additional parameters passed to :class:`~sagemaker.model.Model` |
| |
| .. tip:: |
| |
| You can find additional parameters for using this method at |
| :class:`~sagemaker.model.Model`. |
| |
| Returns: |
| (sagemaker.model.Model) a Model ready for deployment. |
| """ |
| removed_kwargs("serializer", kwargs) |
| removed_kwargs("deserializer", kwargs) |
| removed_kwargs("content_type", kwargs) |
| removed_kwargs("accept", kwargs) |
|
|
| if predictor_cls is None: |
|
|
| def predict_wrapper(endpoint, session): |
| return Predictor(endpoint, session) |
|
|
| predictor_cls = predict_wrapper |
|
|
| role = role or self.role |
|
|
| if "enable_network_isolation" not in kwargs: |
| kwargs["enable_network_isolation"] = self.enable_network_isolation() |
|
|
| return Model( |
| image_uri or self.training_image_uri(), |
| self.model_data, |
| role, |
| vpc_config=self.get_vpc_config(vpc_config_override), |
| sagemaker_session=self.sagemaker_session, |
| predictor_cls=predictor_cls, |
| **kwargs, |
| ) |
|
|
|
|
| class Framework(EstimatorBase): |
| """Base class that cannot be instantiated directly. |
| |
| Subclasses define functionality pertaining to specific ML frameworks, |
| such as training/deployment images and predictor instances. |
| """ |
|
|
| _framework_name = None |
|
|
| def __init__( |
| self, |
| entry_point: Union[str, PipelineVariable], |
| source_dir: Optional[Union[str, PipelineVariable]] = None, |
| hyperparameters: Optional[Dict[str, Union[str, PipelineVariable]]] = None, |
| container_log_level: Union[int, PipelineVariable] = logging.INFO, |
| code_location: Optional[str] = None, |
| image_uri: Optional[Union[str, PipelineVariable]] = None, |
| dependencies: Optional[List[str]] = None, |
| enable_network_isolation: Union[bool, PipelineVariable] = False, |
| git_config: Optional[Dict[str, str]] = None, |
| checkpoint_s3_uri: Optional[Union[str, PipelineVariable]] = None, |
| checkpoint_local_path: Optional[Union[str, PipelineVariable]] = None, |
| enable_sagemaker_metrics: Optional[Union[bool, PipelineVariable]] = None, |
| **kwargs, |
| ): |
| """Base class initializer. |
| |
| Subclasses which override ``__init__`` should invoke ``super()``. |
| |
| Args: |
| entry_point (str or PipelineVariable): Path (absolute or relative) to |
| the local Python source file which should be executed as the entry point to |
| training. If ``source_dir`` is specified, then ``entry_point`` |
| must point to a file located at the root of ``source_dir``. |
| If 'git_config' is provided, 'entry_point' should be |
| a relative location to the Python source file in the Git repo. |
| |
| Example: |
| With the following GitHub repo directory structure: |
| |
| >>> |----- README.md |
| >>> |----- src |
| >>> |----- train.py |
| >>> |----- test.py |
| |
| You can assign entry_point='src/train.py'. |
| source_dir (str or PipelineVariable): Path (absolute, relative or an S3 URI) |
| to a directory with any other training source code dependencies aside from |
| the entry point file (default: None). If ``source_dir`` is an S3 URI, it must |
| point to a tar.gz file. Structure within this directory are preserved |
| when training on Amazon SageMaker. If 'git_config' is provided, |
| 'source_dir' should be a relative location to a directory in the Git |
| repo. |
| |
| .. admonition:: Example |
| |
| With the following GitHub repo directory structure: |
| |
| >>> |----- README.md |
| >>> |----- src |
| >>> |----- train.py |
| >>> |----- test.py |
| |
| and you need 'train.py' as entry point and 'test.py' as |
| training source code as well, you can assign |
| entry_point='train.py', source_dir='src'. |
| hyperparameters (dict[str, str] or dict[str, PipelineVariable]): Hyperparameters |
| that will be used for training (default: None). The hyperparameters are made |
| accessible as a dict[str, str] to the training code on |
| SageMaker. For convenience, this accepts other types for keys |
| and values, but ``str()`` will be called to convert them before |
| training. |
| container_log_level (int or PipelineVariable): Log level to use within the container |
| (default: logging.INFO). Valid values are defined in the Python |
| logging module. |
| code_location (str): The S3 prefix URI where custom code will be |
| uploaded (default: None) - don't include a trailing slash since |
| a string prepended with a "/" is appended to ``code_location``. The code |
| file uploaded to S3 is 'code_location/job-name/source/sourcedir.tar.gz'. |
| If not specified, the default ``code location`` is s3://output_bucket/job-name/. |
| image_uri (str or PipelineVariable): An alternate image name to use instead of the |
| official Sagemaker image for the framework. This is useful to |
| run one of the Sagemaker supported frameworks with an image |
| containing custom dependencies. |
| dependencies (list[str]): A list of paths to directories (absolute |
| or relative) with any additional libraries that will be exported |
| to the container (default: []). The library folders will be |
| copied to SageMaker in the same folder where the entrypoint is |
| copied. If 'git_config' is provided, 'dependencies' should be a |
| list of relative locations to directories with any additional |
| libraries needed in the Git repo. |
| |
| .. admonition:: Example |
| |
| The following call |
| |
| >>> Estimator(entry_point='train.py', |
| ... dependencies=['my/libs/common', 'virtual-env']) |
| |
| results in the following inside the container: |
| |
| >>> $ ls |
| |
| >>> opt/ml/code |
| >>> |------ train.py |
| >>> |------ common |
| >>> |------ virtual-env |
| |
| This is not supported with "local code" in Local Mode. |
| enable_network_isolation (bool or PipelineVariable): Specifies whether container will |
| run in network isolation mode. Network isolation mode restricts |
| the container access to outside networks (such as the internet). |
| The container does not make any inbound or outbound network |
| calls. If True, a channel named "code" will be created for any |
| user entry script for training. The user entry script, files in |
| source_dir (if specified), and dependencies will be uploaded in |
| a tar to S3. Also known as internet-free mode (default: `False`). |
| git_config (dict[str, str]): Git configurations used for cloning |
| files, including ``repo``, ``branch``, ``commit``, |
| ``2FA_enabled``, ``username``, ``password`` and ``token``. The |
| ``repo`` field is required. All other fields are optional. |
| ``repo`` specifies the Git repository where your training script |
| is stored. If you don't provide ``branch``, the default value |
| 'master' is used. If you don't provide ``commit``, the latest |
| commit in the specified branch is used. .. admonition:: Example |
| |
| The following config: |
| |
| >>> git_config = {'repo': 'https://github.com/aws/sagemaker-python-sdk.git', |
| >>> 'branch': 'test-branch-git-config', |
| >>> 'commit': '329bfcf884482002c05ff7f44f62599ebc9f445a'} |
| |
| results in cloning the repo specified in 'repo', then |
| checkout the 'master' branch, and checkout the specified |
| commit. |
| |
| ``2FA_enabled``, ``username``, ``password`` and ``token`` are |
| used for authentication. For GitHub (or other Git) accounts, set |
| ``2FA_enabled`` to 'True' if two-factor authentication is |
| enabled for the account, otherwise set it to 'False'. If you do |
| not provide a value for ``2FA_enabled``, a default value of |
| 'False' is used. CodeCommit does not support two-factor |
| authentication, so do not provide "2FA_enabled" with CodeCommit |
| repositories. |
| |
| For GitHub and other Git repos, when SSH URLs are provided, it |
| doesn't matter whether 2FA is enabled or disabled; you should |
| either have no passphrase for the SSH key pairs, or have the |
| ssh-agent configured so that you will not be prompted for SSH |
| passphrase when you do 'git clone' command with SSH URLs. When |
| HTTPS URLs are provided: if 2FA is disabled, then either token |
| or username+password will be used for authentication if provided |
| (token prioritized); if 2FA is enabled, only token will be used |
| for authentication if provided. If required authentication info |
| is not provided, python SDK will try to use local credentials |
| storage to authenticate. If that fails either, an error message |
| will be thrown. |
| |
| For CodeCommit repos, 2FA is not supported, so '2FA_enabled' |
| should not be provided. There is no token in CodeCommit, so |
| 'token' should not be provided too. When 'repo' is an SSH URL, |
| the requirements are the same as GitHub-like repos. When 'repo' |
| is an HTTPS URL, username+password will be used for |
| authentication if they are provided; otherwise, python SDK will |
| try to use either CodeCommit credential helper or local |
| credential storage for authentication. |
| checkpoint_s3_uri (str or PipelineVariable): The S3 URI in which to persist checkpoints |
| that the algorithm persists (if any) during training. (default: |
| None). |
| checkpoint_local_path (str or PipelineVariable): The local path that the algorithm |
| writes its checkpoints to. SageMaker will persist all files |
| under this path to `checkpoint_s3_uri` continually during |
| training. On job startup the reverse happens - data from the |
| s3 location is downloaded to this path before the algorithm is |
| started. If the path is unset then SageMaker assumes the |
| checkpoints will be provided under `/opt/ml/checkpoints/`. |
| (default: None). |
| enable_sagemaker_metrics (bool or PipelineVariable): enable SageMaker Metrics Time |
| Series. For more information see: |
| https://docs.aws.amazon.com/sagemaker/latest/dg/API_AlgorithmSpecification.html#SageMaker-Type-AlgorithmSpecification-EnableSageMakerMetricsTimeSeries |
| (default: None). |
| **kwargs: Additional kwargs passed to the ``EstimatorBase`` |
| constructor. |
| |
| .. tip:: |
| |
| You can find additional parameters for initializing this class at |
| :class:`~sagemaker.estimator.EstimatorBase`. |
| """ |
| super(Framework, self).__init__(enable_network_isolation=enable_network_isolation, **kwargs) |
| image_uri = renamed_kwargs("image_name", "image_uri", image_uri, kwargs) |
|
|
| validate_source_code_input_against_pipeline_variables( |
| entry_point=entry_point, |
| source_dir=source_dir, |
| git_config=git_config, |
| enable_network_isolation=enable_network_isolation, |
| ) |
| if not is_pipeline_variable(entry_point) and entry_point.startswith("s3://"): |
| raise ValueError( |
| "Invalid entry point script: {}. Must be a path to a local file.".format( |
| entry_point |
| ) |
| ) |
| self.entry_point = entry_point |
| self.git_config = git_config |
| self.source_dir = source_dir |
| self.dependencies = dependencies or [] |
| self.uploaded_code = None |
|
|
| self.container_log_level = container_log_level |
| self.code_location = code_location |
| self.image_uri = image_uri |
|
|
| self._hyperparameters = hyperparameters or {} |
| self.checkpoint_s3_uri = checkpoint_s3_uri |
| self.checkpoint_local_path = checkpoint_local_path |
| self.enable_sagemaker_metrics = enable_sagemaker_metrics |
|
|
| def _prepare_for_training(self, job_name=None): |
| """Set hyperparameters needed for training. This method will also validate ``source_dir``. |
| |
| Args: |
| * job_name (str): Name of the training job to be created. If not |
| specified, one is generated, using the base name given to the |
| constructor if applicable. |
| """ |
| super(Framework, self)._prepare_for_training(job_name=job_name) |
|
|
| self._validate_and_set_debugger_configs() |
|
|
| def _script_mode_hyperparam_update(self, code_dir: str, script: str) -> None: |
| """Applies in-place updates to hyperparameters required for script mode with training. |
| |
| Args: |
| code_dir (str): The directory hosting the training scripts. |
| script (str): The relative filepath of the training entry-point script. |
| """ |
| hyperparams: Dict[str, str] = {} |
| hyperparams[DIR_PARAM_NAME] = code_dir |
| hyperparams[SCRIPT_PARAM_NAME] = script |
| hyperparams[CONTAINER_LOG_LEVEL_PARAM_NAME] = self.container_log_level |
| hyperparams[JOB_NAME_PARAM_NAME] = self._current_job_name |
| hyperparams[SAGEMAKER_REGION_PARAM_NAME] = self.sagemaker_session.boto_region_name |
|
|
| self._hyperparameters.update(hyperparams) |
|
|
| def _validate_and_set_debugger_configs(self): |
| """Set defaults for debugging.""" |
| super(Framework, self)._validate_and_set_debugger_configs() |
|
|
| |
| if self.checkpoint_s3_uri and self.checkpoint_local_path and self.debugger_hook_config: |
| if self._framework_name in {"mxnet", "pytorch", "tensorflow"}: |
| if is_pipeline_variable(self.instance_count): |
| logger.warning( |
| "SMDebug does not currently support distributed training jobs " |
| "with checkpointing enabled. Therefore, to allow parameterized " |
| "instance_count and allow to change it to any values in execution time, " |
| "the debugger_hook_config is disabled." |
| ) |
| self.debugger_hook_config = False |
| elif self.instance_count > 1 or ( |
| hasattr(self, "distribution") |
| and self.distribution is not None |
| ): |
| logger.info( |
| "SMDebug Does Not Currently Support \ |
| Distributed Training Jobs With Checkpointing Enabled" |
| ) |
| self.debugger_hook_config = False |
|
|
| def _model_source_dir(self): |
| """Get the appropriate value to pass as ``source_dir`` to a model constructor. |
| |
| Returns: |
| str: Either a local or an S3 path pointing to the ``source_dir`` to be |
| used for code by the model to be deployed |
| """ |
| if self.sagemaker_session.local_mode: |
| return self.source_dir |
|
|
| if self.uploaded_code is not None: |
| return self.uploaded_code.s3_prefix |
|
|
| return None |
|
|
| def _model_entry_point(self): |
| """Get the appropriate value to pass as ``entry_point`` to a model constructor. |
| |
| Returns: |
| str: The path to the entry point script. This can be either an absolute path or |
| a path relative to ``self._model_source_dir()``. |
| """ |
| if self.sagemaker_session.local_mode or (self._model_source_dir() is None): |
| return self.entry_point |
|
|
| if self.uploaded_code is not None: |
| return self.uploaded_code.script_name |
|
|
| return None |
|
|
| def set_hyperparameters(self, **kwargs): |
| """Escapes the dict argument as JSON, updates the private hyperparameter attribute.""" |
| self._hyperparameters.update(EstimatorBase._json_encode_hyperparameters(kwargs)) |
|
|
| def hyperparameters(self): |
| """Returns the hyperparameters as a dictionary to use for training. |
| |
| The :meth:`~sagemaker.estimator.EstimatorBase.fit` method, which |
| trains the model, calls this method to find the hyperparameters. |
| |
| Returns: |
| dict[str, str]: The hyperparameters. |
| """ |
| return EstimatorBase._json_encode_hyperparameters(self._hyperparameters) |
|
|
| @classmethod |
| def _prepare_init_params_from_job_description(cls, job_details, model_channel_name=None): |
| """Convert the job description to init params that can be handled by the class constructor. |
| |
| Args: |
| job_details: the returned job details from a describe_training_job |
| API call. |
| model_channel_name (str): Name of the channel where pre-trained |
| model data will be downloaded |
| |
| Returns: |
| dictionary: The transformed init_params |
| """ |
| init_params = super(Framework, cls)._prepare_init_params_from_job_description( |
| job_details, model_channel_name |
| ) |
|
|
| init_params["entry_point"] = json.loads( |
| init_params["hyperparameters"].get(SCRIPT_PARAM_NAME) |
| ) |
| init_params["source_dir"] = json.loads(init_params["hyperparameters"].get(DIR_PARAM_NAME)) |
| init_params["container_log_level"] = json.loads( |
| init_params["hyperparameters"].get(CONTAINER_LOG_LEVEL_PARAM_NAME) |
| ) |
|
|
| hyperparameters = {} |
| for k, v in init_params["hyperparameters"].items(): |
| |
| if k == "_tuning_objective_metric": |
| if v.startswith('"') and v.endswith('"'): |
| v = v.strip('"') |
| hyperparameters[k] = v |
| else: |
| hyperparameters[k] = json.loads(v) |
|
|
| init_params["hyperparameters"] = hyperparameters |
|
|
| return init_params |
|
|
| def training_image_uri(self, region=None): |
| """Return the Docker image to use for training. |
| |
| The :meth:`~sagemaker.estimator.EstimatorBase.fit` method, which does |
| the model training, calls this method to find the image to use for model |
| training. |
| |
| Args: |
| region (str): Optional. The AWS Region to use for image URI. Default: AWS Region |
| associated with the SageMaker session. |
| |
| Returns: |
| str: The URI of the Docker image. |
| """ |
|
|
| return image_uris.get_training_image_uri( |
| region=region or self.sagemaker_session.boto_region_name, |
| framework=self._framework_name, |
| framework_version=self.framework_version, |
| py_version=self.py_version, |
| image_uri=self.image_uri, |
| distribution=getattr(self, "distribution", None), |
| compiler_config=getattr(self, "compiler_config", None), |
| tensorflow_version=getattr(self, "tensorflow_version", None), |
| pytorch_version=getattr(self, "pytorch_version", None), |
| instance_type=self._get_instance_type(), |
| ) |
|
|
| @classmethod |
| def attach(cls, training_job_name, sagemaker_session=None, model_channel_name="model"): |
| """Attach to an existing training job. |
| |
| Create an Estimator bound to an existing training job, each subclass |
| is responsible to implement |
| ``_prepare_init_params_from_job_description()`` as this method delegates |
| the actual conversion of a training job description to the arguments |
| that the class constructor expects. After attaching, if the training job |
| has a Complete status, it can be ``deploy()`` ed to create a SageMaker |
| Endpoint and return a ``Predictor``. |
| |
| If the training job is in progress, attach will block until the training job |
| completes, but logs of the training job will not display. To see the logs |
| content, please call ``logs()`` |
| |
| Examples: |
| >>> my_estimator.fit(wait=False) |
| >>> training_job_name = my_estimator.latest_training_job.name |
| Later on: |
| >>> attached_estimator = Estimator.attach(training_job_name) |
| >>> attached_estimator.logs() |
| >>> attached_estimator.deploy() |
| |
| Args: |
| training_job_name (str): The name of the training job to attach to. |
| sagemaker_session (sagemaker.session.Session): Session object which |
| manages interactions with Amazon SageMaker APIs and any other |
| AWS services needed. If not specified, the estimator creates one |
| using the default AWS configuration chain. |
| model_channel_name (str): Name of the channel where pre-trained |
| model data will be downloaded (default: 'model'). If no channel |
| with the same name exists in the training job, this option will |
| be ignored. |
| |
| Returns: |
| Instance of the calling ``Estimator`` Class with the attached |
| training job. |
| """ |
| estimator = super(Framework, cls).attach( |
| training_job_name, sagemaker_session, model_channel_name |
| ) |
|
|
| |
| |
| estimator.uploaded_code = UploadedCode( |
| estimator.source_dir, estimator.entry_point |
| ) |
| return estimator |
|
|
| @classmethod |
| def _update_init_params(cls, hp, tf_arguments): |
| """Placeholder docstring""" |
| updated_params = {} |
| for argument in tf_arguments: |
| value = hp.pop(argument, None) |
| if value is not None: |
| value = json.loads(value) |
| updated_params[argument] = value |
| return updated_params |
|
|
| def transformer( |
| self, |
| instance_count, |
| instance_type, |
| strategy=None, |
| assemble_with=None, |
| output_path=None, |
| output_kms_key=None, |
| accept=None, |
| env=None, |
| max_concurrent_transforms=None, |
| max_payload=None, |
| tags=None, |
| role=None, |
| model_server_workers=None, |
| volume_kms_key=None, |
| entry_point=None, |
| vpc_config_override=vpc_utils.VPC_CONFIG_DEFAULT, |
| enable_network_isolation=None, |
| model_name=None, |
| ): |
| """Return a ``Transformer`` that uses a SageMaker Model based on the training job. |
| |
| It reuses the SageMaker Session and base job name used by |
| the Estimator. |
| |
| Args: |
| instance_count (int): Number of EC2 instances to use. |
| instance_type (str): Type of EC2 instance to use, for example, |
| 'ml.c4.xlarge'. |
| strategy (str): The strategy used to decide how to batch records in |
| a single request (default: None). Valid values: 'MultiRecord' |
| and 'SingleRecord'. |
| assemble_with (str): How the output is assembled (default: None). |
| Valid values: 'Line' or 'None'. |
| output_path (str): S3 location for saving the transform result. If |
| not specified, results are stored to a default bucket. |
| output_kms_key (str): Optional. KMS key ID for encrypting the |
| transform output (default: None). |
| accept (str): The accept header passed by the client to |
| the inference endpoint. If it is supported by the endpoint, |
| it will be the format of the batch transform output. |
| env (dict): Environment variables to be set for use during the |
| transform job (default: None). |
| max_concurrent_transforms (int): The maximum number of HTTP requests |
| to be made to each individual transform container at one time. |
| max_payload (int): Maximum size of the payload in a single HTTP |
| request to the container in MB. |
| tags (list[dict]): List of tags for labeling a transform job. If |
| none specified, then the tags used for the training job are used |
| for the transform job. |
| role (str): The ``ExecutionRoleArn`` IAM Role ARN for the ``Model``, |
| which is also used during transform jobs. If not specified, the |
| role from the Estimator will be used. |
| model_server_workers (int): Optional. The number of worker processes |
| used by the inference server. If None, server will use one |
| worker per vCPU. |
| volume_kms_key (str): Optional. KMS key ID for encrypting the volume |
| attached to the ML compute instance (default: None). |
| entry_point (str): Path (absolute or relative) to the local Python source file which |
| should be executed as the entry point to training. If ``source_dir`` is specified, |
| then ``entry_point`` must point to a file located at the root of ``source_dir``. |
| If not specified, the training entry point is used. |
| vpc_config_override (dict[str, list[str]]): Optional override for |
| the VpcConfig set on the model. |
| Default: use subnets and security groups from this Estimator. |
| |
| * 'Subnets' (list[str]): List of subnet ids. |
| * 'SecurityGroupIds' (list[str]): List of security group ids. |
| |
| enable_network_isolation (bool): Specifies whether container will |
| run in network isolation mode. Network isolation mode restricts |
| the container access to outside networks (such as the internet). |
| The container does not make any inbound or outbound network |
| calls. If True, a channel named "code" will be created for any |
| user entry script for inference. Also known as Internet-free mode. |
| If not specified, this setting is taken from the estimator's |
| current configuration. |
| model_name (str): Name to use for creating an Amazon SageMaker |
| model. If not specified, the estimator generates a default job name |
| based on the training image name and current timestamp. |
| |
| Returns: |
| sagemaker.transformer.Transformer: a ``Transformer`` object that can be used to start a |
| SageMaker Batch Transform job. |
| """ |
| role = role or self.role |
| tags = tags or self.tags |
| model_name = self._get_or_create_name(model_name) |
|
|
| if self.latest_training_job is not None: |
| if enable_network_isolation is None: |
| enable_network_isolation = self.enable_network_isolation() |
|
|
| model = self.create_model( |
| role=role, |
| model_server_workers=model_server_workers, |
| entry_point=entry_point, |
| vpc_config_override=vpc_config_override, |
| model_kms_key=self.output_kms_key, |
| enable_network_isolation=enable_network_isolation, |
| name=model_name, |
| ) |
| model._create_sagemaker_model(instance_type, tags=tags) |
|
|
| transform_env = model.env.copy() |
| if env is not None: |
| transform_env.update(env) |
| else: |
| logger.warning( |
| "No finished training job found associated with this estimator. Please make sure " |
| "this estimator is only used for building workflow config" |
| ) |
| transform_env = env or {} |
|
|
| return Transformer( |
| model_name, |
| instance_count, |
| instance_type, |
| strategy=strategy, |
| assemble_with=assemble_with, |
| output_path=output_path, |
| output_kms_key=output_kms_key, |
| accept=accept, |
| max_concurrent_transforms=max_concurrent_transforms, |
| max_payload=max_payload, |
| env=transform_env, |
| tags=tags, |
| base_transform_job_name=self.base_job_name, |
| volume_kms_key=volume_kms_key, |
| sagemaker_session=self.sagemaker_session, |
| ) |
|
|
| def _distribution_configuration(self, distribution): |
| """Returns a dict of distribution configurations. |
| |
| Args: |
| distribution (dict): A dictionary with information on how to run distributed training. |
| |
| Returns: |
| dict that |
| """ |
| distribution_config = {} |
|
|
| mpi_enabled = False |
| smdataparallel_enabled = False |
| if "instance_groups" in distribution: |
| distribution_config["sagemaker_distribution_instance_groups"] = distribution[ |
| "instance_groups" |
| ] |
|
|
| if "pytorchxla" in distribution: |
| pt_xla_enabled = distribution.get("pytorchxla").get("enabled", False) |
| distribution_config[self.LAUNCH_PT_XLA_ENV_NAME] = pt_xla_enabled |
|
|
| if "parameter_server" in distribution: |
| ps_enabled = distribution.get("parameter_server").get("enabled", False) |
| distribution_config[self.LAUNCH_PS_ENV_NAME] = ps_enabled |
|
|
| if "mpi" in distribution: |
| mpi_dict = distribution["mpi"] |
| mpi_enabled = mpi_dict.get("enabled", False) |
| distribution_config[self.LAUNCH_MPI_ENV_NAME] = mpi_enabled |
|
|
| if mpi_dict.get("processes_per_host"): |
| distribution_config[self.MPI_NUM_PROCESSES_PER_HOST] = mpi_dict.get( |
| "processes_per_host" |
| ) |
|
|
| distribution_config[self.MPI_CUSTOM_MPI_OPTIONS] = mpi_dict.get( |
| "custom_mpi_options", "" |
| ) |
|
|
| if get_mp_parameters(distribution): |
| distribution_config["mp_parameters"] = get_mp_parameters(distribution) |
|
|
| elif "modelparallel" in distribution.get("smdistributed", {}): |
| raise ValueError("Cannot use Model Parallelism without MPI enabled!") |
|
|
| if "smdistributed" in distribution: |
| |
| smdistributed = distribution["smdistributed"] |
| smdataparallel_enabled = smdistributed.get("dataparallel", {}).get("enabled", False) |
| distribution_config[self.LAUNCH_SM_DDP_ENV_NAME] = smdataparallel_enabled |
| distribution_config[self.INSTANCE_TYPE] = self.instance_type |
| if smdataparallel_enabled: |
| distribution_config[self.SM_DDP_CUSTOM_MPI_OPTIONS] = smdistributed[ |
| "dataparallel" |
| ].get("custom_mpi_options", "") |
|
|
| if not (mpi_enabled or smdataparallel_enabled) and distribution_config.get( |
| "sagemaker_distribution_instance_groups" |
| ) not in [None, []]: |
| raise ValueError( |
| "Don't set training instance groups while no distribution strategies enabled!" |
| ) |
|
|
| return distribution_config |
|
|
|
|
| def _s3_uri_prefix(channel_name, s3_data): |
| """Placeholder docstring""" |
| if isinstance(s3_data, TrainingInput): |
| s3_uri = s3_data.config["DataSource"]["S3DataSource"]["S3Uri"] |
| else: |
| s3_uri = s3_data |
| if not s3_uri.startswith("s3://"): |
| raise ValueError("Expecting an s3 uri. Got {}".format(s3_uri)) |
| return {channel_name: s3_uri[5:]} |
|
|
|
|
| |
| |
| def _s3_uri_without_prefix_from_input(input_data): |
| |
| """Placeholder docstring""" |
| if isinstance(input_data, dict): |
| response = {} |
| for channel_name, channel_s3_uri in input_data.items(): |
| response.update(_s3_uri_prefix(channel_name, channel_s3_uri)) |
| return response |
| if isinstance(input_data, str): |
| return _s3_uri_prefix("training", input_data) |
| if isinstance(input_data, TrainingInput): |
| return _s3_uri_prefix("training", input_data) |
| raise ValueError( |
| "Unrecognized type for S3 input data config - not str or TrainingInput: {}".format( |
| input_data |
| ) |
| ) |
|
|