Class Model (1.29.0)

Model(model_name:str,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,version:typing.Optional[str]=None,)

Retrieves the model resource and instantiates its representation.

Parameters

NameDescription
model_namestr

Required. A fully-qualified model resource name or model ID. Example: "projects/123/locations/us-central1/models/456" or "456" when project and location are initialized or passed. May optionally contain a version ID or version alias in {model_name}@{version} form. See version arg.

projectstr

Optional project to retrieve model from. If not set, project set in aiplatform.init will be used.

locationstr

Optional location to retrieve model from. If not set, location set in aiplatform.init will be used.

versionstr

Optional. Version ID or version alias. When set, the specified model version will be targeted unless overridden in method calls. When not set, the model with the "default" alias will be targeted unless overridden in method calls. No behavior change if only one version of a model exists.

Properties

container_spec

The specification of the container that is to be used when deployingthis Model. Not present for AutoML Models.

create_time

Time this resource was created.

description

Description of the model.

display_name

Display name of this resource.

encryption_spec

Customer-managed encryption key options for this Vertex AI resource.

If this is set, then all resources created by this Vertex AI resource willbe encrypted with the provided encryption key.

gca_resource

The underlying resource proto representation.

labels

name

Name of this resource.

predict_schemata

The schemata that describe formats of the Model's predictions andexplanations, if available.

preview

Return a Model instance with preview features enabled.

resource_name

Full qualified resource name, without any version ID.

supported_deployment_resources_types

List of deployment resource types accepted for this Model.

When this Model is deployed, its prediction resources are described bytheprediction_resources field of the objects returned byEndpoint.list_models(). Because not all Models support all resourceconfiguration types, the configuration types this Model supports arelisted here.

If no configuration types are listed, the Model cannot bedeployed to anEndpoint and does not support online predictions(Endpoint.predict() orEndpoint.explain()). Such a Model can servepredictions by using aBatchPredictionJob, if it has at least one entryeach inModel.supported_input_storage_formats andModel.supported_output_storage_formats.

supported_export_formats

The formats and content types in which this Model may be exported.If empty, this Model is not available for export.

For example, if this model can be exported as a Tensorflow SavedModel andhave the artifacts written to Cloud Storage, the expected value would be:

{'tf-saved-model': [<ExportableContent.ARTIFACT: 1>]}

supported_input_storage_formats

The formats this Model supports in theinput_config field of aBatchPredictionJob. IfModel.predict_schemata.instance_schema_uriexists, the instances should be given as per that schema.

Read the docs for more on batch prediction formats

If this Model doesn't support any of these formats it means it cannot beused with aBatchPredictionJob. However, if it hassupported_deployment_resources_types, it could serve online predictionsby usingEndpoint.predict() orEndpoint.explain().

supported_output_storage_formats

The formats this Model supports in theoutput_config field of aBatchPredictionJob.

If bothModel.predict_schemata.instance_schema_uri andModel.predict_schemata.prediction_schema_uri exist, the predictionsare returned together with their instances. In other words, theprediction has the original instance data first, followed by the actualprediction content (as per the schema).

Read the docs for more on batch prediction formats

If this Model doesn't support any of these formats it means it cannot beused with aBatchPredictionJob. However, if it hassupported_deployment_resources_types, it could serve online predictionsby usingEndpoint.predict() orEndpoint.explain().

training_job

The TrainingJob that uploaded this Model, if any.

Exceptions
TypeDescription
api_core.exceptions.NotFoundIf the Model's training job resource cannot be found on the Vertex service.

update_time

Time this resource was last updated.

uri

Path to the directory containing the Model artifact and any of itssupporting files. Not present for AutoML Models.

version_aliases

User provided version aliases so that a model version can be referenced viaalias (i.e. projects/{project}/locations/{location}/models/{model_id}@{version_alias}instead of auto-generated version id (i.e.projects/{project}/locations/{location}/models/{model_id}@{version_id}).The format isa-z][a-zA-Z0-9-]{0,126}[a-z0-9] to distinguish fromversion_id. A default version alias will be created for the first versionof the model, and there must be exactly one default version alias for a model.

version_create_time

Timestamp when this version was created.

version_description

The description of this version.

version_id

The version ID of the model.A new version is committed when a new model version is uploaded ortrained under an existing model id. It is an auto-incrementing decimalnumber in string representation.

version_update_time

Timestamp when this version was updated.

versioned_resource_name

The fully-qualified resource name, including the version ID. For example,projects/{project}/locations/{location}/models/{model_id}@{version_id}

versioning_registry

The registry of model versions associated with thisModel instance.

Methods

batch_predict

batch_predict(job_display_name:typing.Optional[str]=None,gcs_source:typing.Optional[typing.Union[str,typing.Sequence[str]]]=None,bigquery_source:typing.Optional[str]=None,instances_format:str="jsonl",gcs_destination_prefix:typing.Optional[str]=None,bigquery_destination_prefix:typing.Optional[str]=None,predictions_format:str="jsonl",model_parameters:typing.Optional[typing.Dict]=None,machine_type:typing.Optional[str]=None,accelerator_type:typing.Optional[str]=None,accelerator_count:typing.Optional[int]=None,starting_replica_count:typing.Optional[int]=None,max_replica_count:typing.Optional[int]=None,generate_explanation:typing.Optional[bool]=False,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,labels:typing.Optional[typing.Dict[str,str]]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,encryption_spec_key_name:typing.Optional[str]=None,sync:bool=True,create_request_timeout:typing.Optional[float]=None,batch_size:typing.Optional[int]=None,service_account:typing.Optional[str]=None,)->google.cloud.aiplatform.jobs.BatchPredictionJob

Creates a batch prediction job using this Model and outputsprediction results to the provided destination prefix in the specifiedpredictions_format. One source and one destination prefix arerequired.

Example usage: my_model.batch_predict( job_display_name="prediction-123", gcs_source="gs://example-bucket/instances.csv", instances_format="csv", bigquery_destination_prefix="projectId.bqDatasetId.bqTableId" )

Parameters
NameDescription
job_display_namestr

Optional. The user-defined name of the BatchPredictionJob. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

generate_explanationbool

Optional. Generate explanation along with the batch prediction results. This will cause the batch prediction output to include explanations based on theprediction_format: -bigquery: output includes a column namedexplanation. The value is a struct that conforms to the [aiplatform.gapic.Explanation] object. -jsonl: The JSON objects on each line include an additional entry keyedexplanation. The value of the entry is a JSON object that conforms to the [aiplatform.gapic.Explanation] object. -csv: Generating explanations for CSV format is not supported.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Explanation metadata configuration for this BatchPredictionJob. Can be specified only ifgenerate_explanation is set toTrue. This value overrides the value ofModel.explanation_metadata. All fields ofexplanation_metadata are optional in the request. If a field of theexplanation_metadata object is not populated, the corresponding field of theModel.explanation_metadata object is inherited. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. Can be specified only ifgenerate_explanation is set toTrue. This value overrides the value ofModel.explanation_parameters. All fields ofexplanation_parameters are optional in the request. If a field of theexplanation_parameters object is not populated, the corresponding field of theModel.explanation_parameters object is inherited. For more details, seeRef docshttp://tinyurl.com/1an4zake

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

create_request_timeoutfloat

Optional. The timeout for the create request in seconds.

batch_sizeint

Optional. The number of the records (e.g. instances) of the operation given in each batch to a machine replica. Machine type, and size of a single record should be considered when setting this parameter, higher value speeds up the batch operation's execution, but too high value will result in a whole batch not fitting in a machine's memory, and the whole operation will fail. The default value is 64.

service_accountstr

Optional. Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account.

Returns
TypeDescription
job (jobs.BatchPredictionJob)Instantiated representation of the created batch prediction job.

copy

copy(destination_location:str,destination_model_id:typing.Optional[str]=None,destination_parent_model:typing.Optional[str]=None,encryption_spec_key_name:typing.Optional[str]=None,copy_request_timeout:typing.Optional[float]=None,)->google.cloud.aiplatform.models.Model

Copys a model and returns a Model representing the copied Modelresource. This method is a blocking call.

Example usage: copied_model = my_model.copy( destination_location="us-central1" )

Parameters
NameDescription
destination_locationstr

The destination location to copy the model to.

destination_model_idstr

Optional. The ID to use for the copied Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are[a-z0-9_-]. The first character cannot be a number or hyphen. Only set this field when copying as a new model. If this field is not set, a numeric model id will be generated.

destination_parent_modelstr

Optional. The resource name or model ID of an existing model that the newly-copied model will be a version of. Only set this field when copying as a new version of an existing model.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

copy_request_timeoutfloat

Optional. The timeout for the copy request in seconds.

Exceptions
TypeDescription
ValueErrorIf bothdestination_model_id anddestination_parent_model are set.
Returns
TypeDescription
model (aiplatform.Model)Instantiated representation of the copied model resource.

delete

delete(sync:bool=True)->None

Deletes this Vertex AI resource. WARNING: This deletion ispermanent.

Parameter
NameDescription
syncbool

Whether to execute this deletion synchronously. If False, this method will be executed in concurrent Future and any downstream object will be immediately returned and synced when the Future has completed.

deploy

deploy(endpoint:typing.Optional[typing.Union[google.cloud.aiplatform.models.Endpoint,google.cloud.aiplatform.models.PrivateEndpoint,]]=None,deployed_model_display_name:typing.Optional[str]=None,traffic_percentage:typing.Optional[int]=0,traffic_split:typing.Optional[typing.Dict[str,int]]=None,machine_type:typing.Optional[str]=None,min_replica_count:int=1,max_replica_count:int=1,accelerator_type:typing.Optional[str]=None,accelerator_count:typing.Optional[int]=None,service_account:typing.Optional[str]=None,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,metadata:typing.Optional[typing.Sequence[typing.Tuple[str,str]]]=(),encryption_spec_key_name:typing.Optional[str]=None,network:typing.Optional[str]=None,sync=True,deploy_request_timeout:typing.Optional[float]=None,autoscaling_target_cpu_utilization:typing.Optional[int]=None,autoscaling_target_accelerator_duty_cycle:typing.Optional[int]=None,enable_access_logging=False,)->typing.Union[google.cloud.aiplatform.models.Endpoint,google.cloud.aiplatform.models.PrivateEndpoint,]

Deploys model to endpoint. Endpoint will be created if unspecified.

Parameters
NameDescription
endpointUnion[Endpoint, PrivateEndpoint]

Optional. Public or private Endpoint to deploy model to. If not specified, endpoint display name will be model display name+'_endpoint'.

deployed_model_display_namestr

Optional. The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.

traffic_percentageint

Optional. Desired traffic to newly deployed model. Defaults to 0 if there are pre-existing deployed models. Defaults to 100 if there are no pre-existing deployed models. Negative values should not be provided. Traffic of previously deployed models at the endpoint will be scaled down to accommodate new deployed model's traffic. Should not be provided if traffic_split is provided.

traffic_splitDict[str, int]

Optional. A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at the moment. Key for model being deployed is "0". Should not be provided if traffic_percentage is provided.

machine_typestr

Optional. The type of machine. Not specifying machine type will result in model to be deployed with automatic resources.

min_replica_countint

Optional. The minimum number of machine replicas this deployed model will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.

max_replica_countint

Optional. The maximum number of replicas this deployed model may be deployed on when the traffic against it increases. If requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the deployed model increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, the smaller value of min_replica_count or 1 will be used.

accelerator_typestr

Optional. Hardware accelerator type. Must also set accelerator_count if used. One of ACCELERATOR_TYPE_UNSPECIFIED, NVIDIA_TESLA_K80, NVIDIA_TESLA_P100, NVIDIA_TESLA_V100, NVIDIA_TESLA_P4, NVIDIA_TESLA_T4

accelerator_countint

Optional. The number of accelerators to attach to a worker replica.

service_accountstr

The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have theiam.serviceAccounts.actAs permission on this service account.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Metadata describing the Model's input and output for explanation.explanation_metadata is optional whileexplanation_parameters must be specified when used. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. For more details, seeRef docshttp://tinyurl.com/1an4zake

metadataSequence[Tuple[str, str]]

Optional. Strings which should be sent along with the request as metadata.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

networkstr

Optional. The full name of the Compute Engine network to which the Endpoint, if created, will be peered to. E.g. "projects/12345/global/networks/myVPC". Private services access must already be configured for the network. If set or aiplatform.init(network=...) has been set, a PrivateEndpoint will be created. If left unspecified, an Endpoint will be created. Read more about PrivateEndpointsin the documentation.

deploy_request_timeoutfloat

Optional. The timeout for the deploy request in seconds.

autoscaling_target_cpu_utilizationint

Optional. Target CPU Utilization to use for Autoscaling Replicas. A default value of 60 will be used if not specified.

autoscaling_target_accelerator_duty_cycleint

Optional. Target Accelerator Duty Cycle. Must also set accelerator_type and accelerator_count if specified. A default value of 60 will be used if not specified.

syncbool

Whether to execute this method synchronously. If False, this method will be executed in concurrent Future and any downstream object will be immediately returned and synced when the Future has completed.

enable_access_loggingbool

Whether to enable endpoint access logging. Defaults to False.

Exceptions
TypeDescription
ValueErrorIftraffic_split is set for PrivateEndpoint.
Returns
TypeDescription
endpoint (Union[Endpoint, PrivateEndpoint])Endpoint with the deployed model.

export_model

export_model(export_format_id:str,artifact_destination:typing.Optional[str]=None,image_destination:typing.Optional[str]=None,sync:bool=True,)->typing.Dict[str,str]

Exports a trained, exportable Model to a location specified by the user.A Model is considered to be exportable if it has at least onesupported_export_formats.Eitherartifact_destination orimage_destination must be provided.

Example Usage: my_model.export( export_format_id="tf-saved-model", artifact_destination="gs://my-bucket/models/" )

ormy_model.export(    export_format_id="custom-model",    image_destination="us-central1-docker.pkg.dev/projectId/repo/image")
Parameters
NameDescription
export_format_idstr

Required. The ID of the format in which the Model must be exported. The list of export formats that this Model supports can be found by callingModel.supported_export_formats.

artifact_destinationstr

The Cloud Storage location where the Model artifact is to be written to. Under the directory given as the destination a new one with name "model-export-", where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format, will be created. Inside, the Model and any of its supporting files will be written. This field should only be set when, in [Model.supported_export_formats], the value for the key given inexport_format_id containsARTIFACT.

image_destinationstr

The Google Container Registry or Artifact Registry URI where the Model container image will be copied to. Accepted forms: - Google Container Registry path. For example:gcr.io/projectId/imageName:tag. - Artifact Registry path. For example:us-central1-docker.pkg.dev/projectId/repoName/imageName:tag. This field should only be set when, in [Model.supported_export_formats], the value for the key given inexport_format_id containsIMAGE.

syncbool

Whether to execute this export synchronously. If False, this method will be executed in concurrent Future and any downstream object will be immediately returned and synced when the Future has completed.

Exceptions
TypeDescription
ValueErrorIf model does not support exporting.
ValueErrorIf invalid arguments or export formats are provided.
Returns
TypeDescription
output_info (Dict[str, str])Details of the completed export with output destination paths to the artifacts or container image.

get_model_evaluation

get_model_evaluation(evaluation_id:typing.Optional[str]=None,)->typing.Optional[google.cloud.aiplatform.model_evaluation.model_evaluation.ModelEvaluation]

Returns a ModelEvaluation resource and instantiates its representation.If no evaluation_id is passed, it will return the first evaluation associatedwith this model. If the aiplatform.Model resource was instantiated with aversion, this will return a Model Evaluation from that version. If no versionwas specified when instantiating the Model resource, this will return anEvaluation from the default version.

Example usage: my_model = Model( model_name="projects/123/locations/us-central1/models/456" )

my_evaluation = my_model.get_model_evaluation(    evaluation_id="789")# If no arguments are passed, this method returns the first evaluation for the modelmy_evaluation = my_model.get_model_evaluation()
Parameter
NameDescription
evaluation_idstr

Optional. The ID of the model evaluation to retrieve.

Returns
TypeDescription
model_evaluation.ModelEvaluationInstantiated representation of the ModelEvaluation resource.

list

list(filter:typing.Optional[str]=None,order_by:typing.Optional[str]=None,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,)->typing.List[google.cloud.aiplatform.models.Model]

List all Model resource instances.

Example Usage: aiplatform.Model.list( filter='labels.my_label="my_label_value" AND display_name="my_model"', )

Parameters
NameDescription
filterstr

Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported.

order_bystr

Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields:display_name,create_time,update_time

projectstr

Optional. Project to retrieve list from. If not set, project set in aiplatform.init will be used.

locationstr

Optional. Location to retrieve list from. If not set, location set in aiplatform.init will be used.

credentialsauth_credentials.Credentials

Optional. Custom credentials to use to retrieve list. Overrides credentials set in aiplatform.init.

Returns
TypeDescription
List[models.Model]A list of Model resource objects

list_model_evaluations

list_model_evaluations()->(typing.List[google.cloud.aiplatform.model_evaluation.model_evaluation.ModelEvaluation])

List all Model Evaluation resources associated with this model.If this Model resource was instantiated with a version, the ModelEvaluation resources for that version will be returned. If no versionwas provided when the Model resource was instantiated, Model Evaluationresources will be returned for the default version.

Example Usage: my_model = Model( model_name="projects/123/locations/us-central1/models/456@1" )

my_evaluations = my_model.list_model_evaluations()
Returns
TypeDescription
List[model_evaluation.ModelEvaluation]List of ModelEvaluation resources for the model.

to_dict

to_dict()->typing.Dict[str,typing.Any]

Returns the resource proto as a dictionary.

update

update(display_name:typing.Optional[str]=None,description:typing.Optional[str]=None,labels:typing.Optional[typing.Dict[str,str]]=None,)->google.cloud.aiplatform.models.Model

Updates a model.

Example usage: my_model = my_model.update( display_name="my-model", description="my description", labels={'key': 'value'}, )

Parameters
NameDescription
display_namestr

The display name of the Model. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

descriptionstr

The description of the model.

labelsDict[str, str]

Optional. The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. Seehttps://goo.gl/xmQnxf for more information and examples of labels.

Exceptions
TypeDescription
ValueErrorIflabels is not the correct format.
Returns
TypeDescription
model (aiplatform.Model)Updated model resource.

upload

upload(serving_container_image_uri:typing.Optional[str]=None,*,artifact_uri:typing.Optional[str]=None,model_id:typing.Optional[str]=None,parent_model:typing.Optional[str]=None,is_default_version:bool=True,version_aliases:typing.Optional[typing.Sequence[str]]=None,version_description:typing.Optional[str]=None,serving_container_predict_route:typing.Optional[str]=None,serving_container_health_route:typing.Optional[str]=None,description:typing.Optional[str]=None,serving_container_command:typing.Optional[typing.Sequence[str]]=None,serving_container_args:typing.Optional[typing.Sequence[str]]=None,serving_container_environment_variables:typing.Optional[typing.Dict[str,str]]=None,serving_container_ports:typing.Optional[typing.Sequence[int]]=None,local_model:typing.Optional[LocalModel]=None,instance_schema_uri:typing.Optional[str]=None,parameters_schema_uri:typing.Optional[str]=None,prediction_schema_uri:typing.Optional[str]=None,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,display_name:typing.Optional[str]=None,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,labels:typing.Optional[typing.Dict[str,str]]=None,encryption_spec_key_name:typing.Optional[str]=None,staging_bucket:typing.Optional[str]=None,sync=True,upload_request_timeout:typing.Optional[float]=None)->Model

Uploads a model and returns a Model representing the uploaded Modelresource.

Example usage: my_model = Model.upload( display_name="my-model", artifact_uri="gs://my-model/saved-model", serving_container_image_uri="tensorflow/serving" )

Parameters
NameDescription
serving_container_image_uristr

Optional. The URI of the Model serving container. This parameter is required if the parameterlocal_model is not specified.

artifact_uristr

Optional. The path to the directory containing the Model artifact and any of its supporting files. Leave blank for custom container prediction. Not present for AutoML Models.

model_idstr

Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are[a-z0-9_-]. The first character cannot be a number or hyphen.

parent_modelstr

Optional. The resource name or model ID of an existing model that the newly-uploaded model will be a version of. Only set this field when uploading a new version of an existing model.

is_default_versionbool

Optional. When set to True, the newly uploaded model version will automatically have alias "default" included. Subsequent uses of this model without a version specified will use this "default" version. When set to False, the "default" alias will not be moved. Actions targeting the newly-uploaded model version will need to specifically reference this version by ID or alias. New model uploads, i.e. version 1, will always be "default" aliased.

version_aliasesSequence[str]

Optional. User provided version aliases so that a model version can be referenced via alias instead of auto-generated version ID. A default version alias will be created for the first version of the model. The format isa-z][a-zA-Z0-9-]{0,126}[a-z0-9]

version_descriptionstr

Optional. The description of the model version being uploaded.

serving_container_predict_routestr

Optional. An HTTP path to send prediction requests to the container, and which must be supported by it. If not specified a default HTTP path will be used by Vertex AI.

serving_container_health_routestr

Optional. An HTTP path to send health check requests to the container, and which must be supported by it. If not specified a standard HTTP path will be used by Vertex AI.

descriptionstr

The description of the model.

local_modelOptional[LocalModel]

Optional. A LocalModel instance that includes aserving_container_spec. If provided, theserving_container_spec of the LocalModel instance will overwrite the values of all other serving container parameters.

instance_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used inPredictRequest.instances,ExplainRequest.instances andBatchPredictionJob.input_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

parameters_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation viaPredictRequest.parameters,ExplainRequest.parameters andBatchPredictionJob.model_parameters. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform, if no parameters are supported it is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

prediction_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned viaPredictResponse.predictions,ExplainResponse.explanations, andBatchPredictionJob.output_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Metadata describing the Model's input and output for explanation.explanation_metadata is optional whileexplanation_parameters must be specified when used. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. For more details, seeRef docshttp://tinyurl.com/1an4zake

display_namestr

Optional. The display name of the Model. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

labelsDict[str, str]

Optional. The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. Seehttps://goo.gl/xmQnxf for more information and examples of labels.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

staging_bucketstr

Optional. Bucket to stage local model artifacts. Overrides staging_bucket set in aiplatform.init.

upload_request_timeoutfloat

Optional. The timeout for the upload request in seconds.

Exceptions
TypeDescription
ValueErrorIf explanation_metadata is specified while explanation_parameters is not. Also if model directory does not contain a supported model file. Iflocal_model is specified butserving_container_spec.image_uri in thelocal_model is None. Iflocal_model is not specified andserving_container_image_uri is None.
Returns
TypeDescription
model (aiplatform.Model)Instantiated representation of the uploaded model resource.

upload_scikit_learn_model_file

upload_scikit_learn_model_file(model_file_path:str,sklearn_version:typing.Optional[str]=None,display_name:typing.Optional[str]=None,description:typing.Optional[str]=None,model_id:typing.Optional[str]=None,parent_model:typing.Optional[str]=None,is_default_version:typing.Optional[bool]=True,version_aliases:typing.Optional[typing.Sequence[str]]=None,version_description:typing.Optional[str]=None,instance_schema_uri:typing.Optional[str]=None,parameters_schema_uri:typing.Optional[str]=None,prediction_schema_uri:typing.Optional[str]=None,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,labels:typing.Optional[typing.Dict[str,str]]=None,encryption_spec_key_name:typing.Optional[str]=None,staging_bucket:typing.Optional[str]=None,sync=True,upload_request_timeout:typing.Optional[float]=None,)->google.cloud.aiplatform.models.Model

Uploads a model and returns a Model representing the uploaded Modelresource.

Example usage: my_model = Model.upload_scikit_learn_model_file( model_file_path="iris.sklearn_model.joblib" )

Parameters
NameDescription
model_file_pathstr

Required. Local file path of the model.

sklearn_versionstr

Optional. The version of the Scikit-learn serving container. Supported versions: ["0.20", "0.22", "0.23", "0.24", "1.0"]. If the version is not specified, the latest version is used.

display_namestr

Optional. The display name of the Model. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

descriptionstr

The description of the model.

model_idstr

Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are[a-z0-9_-]. The first character cannot be a number or hyphen.

parent_modelstr

Optional. The resource name or model ID of an existing model that the newly-uploaded model will be a version of. Only set this field when uploading a new version of an existing model.

is_default_versionbool

Optional. When set to True, the newly uploaded model version will automatically have alias "default" included. Subsequent uses of this model without a version specified will use this "default" version. When set to False, the "default" alias will not be moved. Actions targeting the newly-uploaded model version will need to specifically reference this version by ID or alias. New model uploads, i.e. version 1, will always be "default" aliased.

version_aliasesSequence[str]

Optional. User provided version aliases so that a model version can be referenced via alias instead of auto-generated version ID. A default version alias will be created for the first version of the model. The format isa-z][a-zA-Z0-9-]{0,126}[a-z0-9]

version_descriptionstr

Optional. The description of the model version being uploaded.

instance_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used inPredictRequest.instances,ExplainRequest.instances andBatchPredictionJob.input_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

parameters_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation viaPredictRequest.parameters,ExplainRequest.parameters andBatchPredictionJob.model_parameters. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform, if no parameters are supported it is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

prediction_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned viaPredictResponse.predictions,ExplainResponse.explanations, andBatchPredictionJob.output_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Metadata describing the Model's input and output for explanation.explanation_metadata is optional whileexplanation_parameters must be specified when used. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. For more details, seeRef docshttp://tinyurl.com/1an4zake

labelsDict[str, str]

Optional. The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. Seehttps://goo.gl/xmQnxf for more information and examples of labels.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

staging_bucketstr

Optional. Bucket to stage local model artifacts. Overrides staging_bucket set in aiplatform.init.

syncbool

Whether to execute this method synchronously. If False, this method will be executed in concurrent Future and any downstream object will be immediately returned and synced when the Future has completed.

upload_request_timeoutfloat

Optional. The timeout for the upload request in seconds.

Exceptions
TypeDescription
ValueErrorIf explanation_metadata is specified while explanation_parameters is not. Also if model directory does not contain a supported model file.
Returns
TypeDescription
model (aiplatform.Model)Instantiated representation of the uploaded model resource.

upload_tensorflow_saved_model

upload_tensorflow_saved_model(saved_model_dir:str,tensorflow_version:typing.Optional[str]=None,use_gpu:bool=False,display_name:typing.Optional[str]=None,description:typing.Optional[str]=None,model_id:typing.Optional[str]=None,parent_model:typing.Optional[str]=None,is_default_version:typing.Optional[bool]=True,version_aliases:typing.Optional[typing.Sequence[str]]=None,version_description:typing.Optional[str]=None,instance_schema_uri:typing.Optional[str]=None,parameters_schema_uri:typing.Optional[str]=None,prediction_schema_uri:typing.Optional[str]=None,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,labels:typing.Optional[typing.Dict[str,str]]=None,encryption_spec_key_name:typing.Optional[str]=None,staging_bucket:typing.Optional[str]=None,sync=True,upload_request_timeout:typing.Optional[str]=None,)->google.cloud.aiplatform.models.Model

Uploads a model and returns a Model representing the uploaded Modelresource.

Example usage: my_model = Model.upload_scikit_learn_model_file( model_file_path="iris.tensorflow_model.SavedModel" )

Parameters
NameDescription
saved_model_dirstr

Required. Local directory of the Tensorflow SavedModel.

tensorflow_versionstr

Optional. The version of the Tensorflow serving container. Supported versions: ["0.15", "2.1", "2.2", "2.3", "2.4", "2.5", "2.6", "2.7"]. If the version is not specified, the latest version is used.

use_gpubool

Whether to use GPU for model serving.

display_namestr

Optional. The display name of the Model. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

descriptionstr

The description of the model.

model_idstr

Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are[a-z0-9_-]. The first character cannot be a number or hyphen.

parent_modelstr

Optional. The resource name or model ID of an existing model that the newly-uploaded model will be a version of. Only set this field when uploading a new version of an existing model.

is_default_versionbool

Optional. When set to True, the newly uploaded model version will automatically have alias "default" included. Subsequent uses of this model without a version specified will use this "default" version. When set to False, the "default" alias will not be moved. Actions targeting the newly-uploaded model version will need to specifically reference this version by ID or alias. New model uploads, i.e. version 1, will always be "default" aliased.

version_aliasesSequence[str]

Optional. User provided version aliases so that a model version can be referenced via alias instead of auto-generated version ID. A default version alias will be created for the first version of the model. The format isa-z][a-zA-Z0-9-]{0,126}[a-z0-9]

version_descriptionstr

Optional. The description of the model version being uploaded.

instance_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used inPredictRequest.instances,ExplainRequest.instances andBatchPredictionJob.input_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

parameters_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation viaPredictRequest.parameters,ExplainRequest.parameters andBatchPredictionJob.model_parameters. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform, if no parameters are supported it is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

prediction_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned viaPredictResponse.predictions,ExplainResponse.explanations, andBatchPredictionJob.output_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Metadata describing the Model's input and output for explanation.explanation_metadata is optional whileexplanation_parameters must be specified when used. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. For more details, seeRef docshttp://tinyurl.com/1an4zake

labelsDict[str, str]

Optional. The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. Seehttps://goo.gl/xmQnxf for more information and examples of labels.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

staging_bucketstr

Optional. Bucket to stage local model artifacts. Overrides staging_bucket set in aiplatform.init.

upload_request_timeoutfloat

Optional. The timeout for the upload request in seconds.

syncbool

Whether to execute this method synchronously. If False, this method will be executed in concurrent Future and any downstream object will be immediately returned and synced when the Future has completed.

Exceptions
TypeDescription
ValueErrorIf explanation_metadata is specified while explanation_parameters is not. Also if model directory does not contain a supported model file.
Returns
TypeDescription
model (aiplatform.Model)Instantiated representation of the uploaded model resource.

upload_xgboost_model_file

upload_xgboost_model_file(model_file_path:str,xgboost_version:typing.Optional[str]=None,display_name:typing.Optional[str]=None,description:typing.Optional[str]=None,model_id:typing.Optional[str]=None,parent_model:typing.Optional[str]=None,is_default_version:typing.Optional[bool]=True,version_aliases:typing.Optional[typing.Sequence[str]]=None,version_description:typing.Optional[str]=None,instance_schema_uri:typing.Optional[str]=None,parameters_schema_uri:typing.Optional[str]=None,prediction_schema_uri:typing.Optional[str]=None,explanation_metadata:typing.Optional[google.cloud.aiplatform_v1.types.explanation_metadata.ExplanationMetadata]=None,explanation_parameters:typing.Optional[google.cloud.aiplatform_v1.types.explanation.ExplanationParameters]=None,project:typing.Optional[str]=None,location:typing.Optional[str]=None,credentials:typing.Optional[google.auth.credentials.Credentials]=None,labels:typing.Optional[typing.Dict[str,str]]=None,encryption_spec_key_name:typing.Optional[str]=None,staging_bucket:typing.Optional[str]=None,sync=True,upload_request_timeout:typing.Optional[float]=None,)->google.cloud.aiplatform.models.Model

Uploads a model and returns a Model representing the uploaded Modelresource.

Example usage: my_model = Model.upload_xgboost_model_file( model_file_path="iris.xgboost_model.bst" )

Parameters
NameDescription
model_file_pathstr

Required. Local file path of the model.

xgboost_versionstr

Optional. The version of the XGBoost serving container. Supported versions: ["0.82", "0.90", "1.1", "1.2", "1.3", "1.4"]. If the version is not specified, the latest version is used.

display_namestr

Optional. The display name of the Model. The name can be up to 128 characters long and can be consist of any UTF-8 characters.

descriptionstr

The description of the model.

model_idstr

Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are[a-z0-9_-]. The first character cannot be a number or hyphen.

parent_modelstr

Optional. The resource name or model ID of an existing model that the newly-uploaded model will be a version of. Only set this field when uploading a new version of an existing model.

is_default_versionbool

Optional. When set to True, the newly uploaded model version will automatically have alias "default" included. Subsequent uses of this model without a version specified will use this "default" version. When set to False, the "default" alias will not be moved. Actions targeting the newly-uploaded model version will need to specifically reference this version by ID or alias. New model uploads, i.e. version 1, will always be "default" aliased.

version_aliasesSequence[str]

Optional. User provided version aliases so that a model version can be referenced via alias instead of auto-generated version ID. A default version alias will be created for the first version of the model. The format isa-z][a-zA-Z0-9-]{0,126}[a-z0-9]

version_descriptionstr

Optional. The description of the model version being uploaded.

instance_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used inPredictRequest.instances,ExplainRequest.instances andBatchPredictionJob.input_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

parameters_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation viaPredictRequest.parameters,ExplainRequest.parameters andBatchPredictionJob.model_parameters. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform, if no parameters are supported it is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

prediction_schema_uristr

Optional. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned viaPredictResponse.predictions,ExplainResponse.explanations, andBatchPredictionJob.output_config. The schema is defined as an OpenAPI 3.0.2Schema Objecthttps://tinyurl.com/y538mdwt#schema-object__. AutoML Models always have this field populated by AI Platform. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.

explanation_metadataaiplatform.explain.ExplanationMetadata

Optional. Metadata describing the Model's input and output for explanation.explanation_metadata is optional whileexplanation_parameters must be specified when used. For more details, seeRef docshttp://tinyurl.com/1igh60kt

explanation_parametersaiplatform.explain.ExplanationParameters

Optional. Parameters to configure explaining for Model's predictions. For more details, seeRef docshttp://tinyurl.com/1an4zake

labelsDict[str, str]

Optional. The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. Seehttps://goo.gl/xmQnxf for more information and examples of labels.

encryption_spec_key_nameOptional[str]

Optional. The Cloud KMS resource identifier of the customer managed encryption key used to protect the model. Has the form:projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key. The key needs to be in the same region as where the compute resource is created. If set, this Model and all sub-resources of this Model will be secured by this key. Overrides encryption_spec_key_name set in aiplatform.init.

staging_bucketstr

Optional. Bucket to stage local model artifacts. Overrides staging_bucket set in aiplatform.init.

upload_request_timeoutfloat

Optional. The timeout for the upload request in seconds.

Exceptions
TypeDescription
ValueErrorIf model directory does not contain a supported model file.
Returns
TypeDescription
model (aiplatform.Model)Instantiated representation of the uploaded model resource.

wait

wait()

Helper method that blocks until all futures are complete.

Except as otherwise noted, the content of this page is licensed under theCreative Commons Attribution 4.0 License, and code samples are licensed under theApache 2.0 License. For details, see theGoogle Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025-10-30 UTC.