| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from typing import Dict, Literal, Optional, Union |
|
|
| from google.cloud.aiplatform.utils import _ipython_utils |
| from google.cloud.aiplatform_v1beta1.types import ( |
| tuning_job as gca_tuning_job_types, |
| ) |
| from vertexai import generative_models |
| from vertexai.tuning import _tuning |
|
|
|
|
| def train( |
| *, |
| source_model: Union[str, generative_models.GenerativeModel], |
| train_dataset: str, |
| validation_dataset: Optional[str] = None, |
| tuned_model_display_name: Optional[str] = None, |
| epochs: Optional[int] = None, |
| learning_rate_multiplier: Optional[float] = None, |
| adapter_size: Optional[Literal[1, 4, 8, 16]] = None, |
| labels: Optional[Dict[str, str]] = None, |
| ) -> "SupervisedTuningJob": |
| """Tunes a model using supervised training. |
| |
| Args: |
| source_model (str): Model name for tuning, e.g., "gemini-1.0-pro-002". |
| train_dataset: Cloud Storage path to file containing training dataset for |
| tuning. The dataset should be in JSONL format. |
| validation_dataset: Cloud Storage path to file containing validation |
| dataset for tuning. The dataset should be in JSONL format. |
| tuned_model_display_name: The display name of the |
| [TunedModel][google.cloud.aiplatform.v1.Model]. The name can be up to |
| 128 characters long and can consist of any UTF-8 characters. |
| epochs: Number of training epoches for this tuning job. |
| learning_rate_multiplier: Learning rate multiplier for tuning. |
| adapter_size: Adapter size for tuning. |
| labels: User-defined metadata to be associated with trained models |
| |
| Returns: |
| A `TuningJob` object. |
| """ |
| if adapter_size is None: |
| adapter_size_value = None |
| elif adapter_size == 1: |
| adapter_size_value = ( |
| gca_tuning_job_types.SupervisedHyperParameters.AdapterSize.ADAPTER_SIZE_ONE |
| ) |
| elif adapter_size == 4: |
| adapter_size_value = ( |
| gca_tuning_job_types.SupervisedHyperParameters.AdapterSize.ADAPTER_SIZE_FOUR |
| ) |
| elif adapter_size == 8: |
| adapter_size_value = ( |
| gca_tuning_job_types.SupervisedHyperParameters.AdapterSize.ADAPTER_SIZE_EIGHT |
| ) |
| elif adapter_size == 16: |
| adapter_size_value = ( |
| gca_tuning_job_types.SupervisedHyperParameters.AdapterSize.ADAPTER_SIZE_SIXTEEN |
| ) |
| else: |
| raise ValueError( |
| f"Unsupported adapter size: {adapter_size}. The supported sizes are [1, 4, 8, 16]" |
| ) |
| supervised_tuning_spec = gca_tuning_job_types.SupervisedTuningSpec( |
| training_dataset_uri=train_dataset, |
| validation_dataset_uri=validation_dataset, |
| hyper_parameters=gca_tuning_job_types.SupervisedHyperParameters( |
| epoch_count=epochs, |
| learning_rate_multiplier=learning_rate_multiplier, |
| adapter_size=adapter_size_value, |
| ), |
| ) |
|
|
| if isinstance(source_model, generative_models.GenerativeModel): |
| source_model = source_model._prediction_resource_name.rpartition("/")[-1] |
|
|
| supervised_tuning_job = ( |
| SupervisedTuningJob._create( |
| base_model=source_model, |
| tuning_spec=supervised_tuning_spec, |
| tuned_model_display_name=tuned_model_display_name, |
| labels=labels, |
| ) |
| ) |
| _ipython_utils.display_model_tuning_button(supervised_tuning_job) |
|
|
| return supervised_tuning_job |
|
|
|
|
| class SupervisedTuningJob(_tuning.TuningJob): |
| def __init__(self, tuning_job_name: str): |
| super().__init__(tuning_job_name=tuning_job_name) |
| _ipython_utils.display_model_tuning_button(self) |
|
|