
                         f    S r SSKJr  SSKJr  SSKJr  SSKJr  SSKJr  SSK	J
r
   " S S	\5      rg
)z%Utilities for AI Platform models API.    )absolute_import)division)unicode_literals)
list_pager)apis)	constantsc                       \ rS rSrSrSS jr                           SS jr                          SS jrS rS r	S	 r
SS
 jrSS jr     SS jr     SS jrSrg)ModelsClient   a  High-level client for the AI Platform models surface.

Attributes:
  client: An instance of the given client, or the API client aiplatform of
    Beta version.
  messages: The messages module for the given client, or the API client
    aiplatform of Beta version.
Nc                 &   U=(       dD    [         R                  " [        R                  [        R                  [        R
                     5      U l        U=(       d    U R                  R                  U l        U R                  R                  U l
        g )N)r   GetClientInstancer   AI_PLATFORM_API_NAMEAI_PLATFORM_API_VERSIONBETA_VERSIONclientMESSAGES_MODULEmessagesprojects_locations_models_service)selfr   r   s      .lib/googlecloudsdk/api_lib/ai/models/client.py__init__ModelsClient.__init__$   sc     CD22&&)))*@*@ACDK ; ; ;DMKK99DM    c                 n   U R                   R                  UUUS9nU(       a  UUl        U(       a  UUl        U	(       a3  U	 Vs/ s H   nU R                   R	                  UU	U   S9PM"     snUl        U
(       a/  U
 Vs/ s H  nU R                   R                  US9PM     snUl        U(       a/  U Vs/ s H  nU R                   R                  US9PM     snUl        U(       a  [        U5      S-   Ul
        U(       a  UUl        U(       d  U(       d  U(       aB  SnU(       a  U R                   R                  US9nU R                   R                  UUUS9Ul        U(       d  U(       d  U(       aB  Sn U(       a  U R                   R                  US9n U R                   R                  U UUS9Ul        U R                   R!                  UUUUUUUS9n!U(       a  UU!l        U(       af  / n"[%        UR'                  5       5       H3  u  n#n$U"R)                  U!R+                  5       R-                  U#U$S	95        M5     U!R+                  U"S
9U!l        U R0                  R3                  U R                   R5                  UR7                  5       U R                   R9                  U!UUS9S95      $ s  snf s  snf s  snf )a8  Constructs, sends an UploadModel request and returns the LRO to be done.

Args:
  region_ref: The resource reference for a given region. None if the region
    reference is not provided.
  display_name: The display name of the Model. The name can be up to 128
    characters long and can be consist of any UTF-8 characters.
  description: The description of the Model.
  version_description: The description of the Model version.
  artifact_uri: The path to the directory containing the Model artifact and
    any of its supporting files. Not present for AutoML Models.
  container_image_uri: Immutable. URI of the Docker image to be used as the
    custom container for serving predictions. This URI must identify an
    image in Artifact Registry or Container Registry. Learn more about the
    [container publishing requirements](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#publishing), including
    permissions requirements for the Vertex AI Service Agent. The container
    image is ingested upon ModelService.UploadModel, stored internally, and
    this original path is afterwards not used. To learn about the
    requirements for the Docker image itself, see [Custom container
    requirements](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#). You can use the URI
    to one of Vertex AI's [pre-built container images for
    prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-
    built-containers) in this field.
  container_command: Specifies the command that runs when the container
    starts. This overrides the container's [ENTRYPOINT](https://docs.docker.
    com/engine/reference/builder/#entrypoint). Specify this field as an
    array of executable and arguments, similar to a Docker `ENTRYPOINT`'s
    "exec" form, not its "shell" form. If you do not specify this field,
    then the container's `ENTRYPOINT` runs, in conjunction with the args
    field or the container's
    [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd), if
    either exists. If this field is not specified and the container does not
    have an `ENTRYPOINT`, then refer to the Docker documentation about [how
    `CMD` and `ENTRYPOINT`
    interact](https://docs.docker.com/engine/reference/builder/#understand-
    how-cmd-and-entrypoint-interact). If you specify this field, then you
    can also specify the `args` field to provide additional arguments for
    this command. However, if you specify this field, then the container's
    `CMD` is ignored. See the [Kubernetes documentation about how the
    `command` and `args` fields interact with a container's `ENTRYPOINT` and
    `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-
    command-argument-container/#notes). In this field, you can reference
    [environment variables set by Vertex
    AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables) and environment variables set in
    the env field. You cannot reference environment variables set in the
    Docker image. In order for environment variables to be expanded,
    reference them by using the following syntax: $( VARIABLE_NAME) Note
    that this differs from Bash variable expansion, which does not use
    parentheses. If a variable cannot be resolved, the reference in the
    input string is used unchanged. To avoid variable expansion, you can
    escape this syntax with `$$`; for example: $$(VARIABLE_NAME) This field
    corresponds to the `command` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_args: Specifies arguments for the command that runs when the
    container starts. This overrides the container's
    [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd). Specify
    this field as an array of executable and arguments, similar to a Docker
    `CMD`'s "default parameters" form. If you don't specify this field but
    do specify the command field, then the command from the `command` field
    runs without any additional arguments. See the [Kubernetes documentation
    about how the `command` and `args` fields interact with a container's
    `ENTRYPOINT` and `CMD`](https://kubernetes.io/docs/tasks/inject-data-
    application/define-command-argument-container/#notes). If you don't
    specify this field and don't specify the `command` field, then the
    container's
    [`ENTRYPOINT`](https://docs.docker.com/engine/reference/builder/#cmd)
    and `CMD` determine what runs based on their default behavior. See the
    Docker documentation about [how `CMD` and `ENTRYPOINT`
    interact](https://docs.docker.com/engine/reference/builder/#understand-
    how-cmd-and-entrypoint-interact). In this field, you can reference
    [environment variables set by Vertex
    AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables) and environment variables set in
    the env field. You cannot reference environment variables set in the
    Docker image. In order for environment variables to be expanded,
    reference them by using the following syntax: $( VARIABLE_NAME) Note
    that this differs from Bash variable expansion, which does not use
    parentheses. If a variable cannot be resolved, the reference in the
    input string is used unchanged. To avoid variable expansion, you can
    escape this syntax with `$$`; for example: $$(VARIABLE_NAME) This field
    corresponds to the `args` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core)..
  container_env_vars: List of environment variables to set in the container.
    After the container starts running, code running in the container can
    read these environment variables. Additionally, the command and args
    fields can reference these variables. Later entries in this list can
    also reference earlier entries. For example, the following example sets
    the variable `VAR_2` to have the value `foo bar`: ```json [ { "name":
    "VAR_1", "value": "foo" }, { "name": "VAR_2", "value": "$(VAR_1) bar" }
    ] ``` If you switch the order of the variables in the example, then the
    expansion does not occur. This field corresponds to the `env` field of
    the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_ports: List of ports to expose from the container. Vertex AI
    sends any http prediction requests that it receives to the first port on
    this list. Vertex AI also sends [liveness and health
    checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#liveness) to this port. If you do not specify
    this field, it defaults to following value: ```json [ { "containerPort":
    8080 } ] ``` Vertex AI does not use ports other than the first one
    listed. This field corresponds to the `ports` field of the Kubernetes
    Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_grpc_ports: List of ports to expose from the container. Vertex
    AI sends any grpc prediction requests that it receives to the first port
    on this list. Vertex AI also sends [liveness and health
    checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#liveness) to this port. If you do not specify
    this field, gRPC requests to the container will be disabled. Vertex AI
    does not use ports other than the first one listed. This field
    corresponds to the `ports` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_predict_route: HTTP path on the container to send prediction
    requests to. Vertex AI forwards requests sent using
    projects.locations.endpoints.predict to this path on the container's IP
    address and port. Vertex AI then returns the container's response in the
    API response. For example, if you set this field to `/foo`, then when
    Vertex AI receives a prediction request, it forwards the request body in
    a POST request to the `/foo` path on the port of your container
    specified by the first value of this `ModelContainerSpec`'s ports field.
    If you don't specify this field, it defaults to the following value when
    you deploy this Model to an Endpoint:
    /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The
    placeholders in this value are replaced as follows: * ENDPOINT: The last
    segment (following `endpoints/`)of the Endpoint.name][] field of the
    Endpoint where this Model has been deployed. (Vertex AI makes this value
    available to your container code as the [`AIP_ENDPOINT_ID` environment
    variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables).) * DEPLOYED_MODEL:
    DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value
    available to your container code as the [`AIP_DEPLOYED_MODEL_ID`
    environment variable](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#aip-variables).)
  container_health_route: HTTP path on the container to send health checks
    to. Vertex AI intermittently sends GET requests to this path on the
    container's IP address and port to check that the container is healthy.
    Read more about [health checks](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#health). For example,
    if you set this field to `/bar`, then Vertex AI intermittently sends a
    GET request to the `/bar` path on the port of your container specified
    by the first value of this `ModelContainerSpec`'s ports field. If you
    don't specify this field, it defaults to the following value when you
    deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/
    DEPLOYED_MODEL:predict The placeholders in this value are replaced as
    follows * ENDPOINT: The last segment (following `endpoints/`)of the
    Endpoint.name][] field of the Endpoint where this Model has been
    deployed. (Vertex AI makes this value available to your container code
    as the [`AIP_ENDPOINT_ID` environment
    variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables).) * DEPLOYED_MODEL:
    DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value
    available to your container code as the [`AIP_DEPLOYED_MODEL_ID`
    environment variable](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#aip-variables).)
  container_deployment_timeout_seconds (int): Deployment timeout in seconds.
  container_shared_memory_size_mb (int): The amount of the VM memory to
    reserve as the shared memory for the model in megabytes.
  container_startup_probe_exec (Sequence[str]): Exec specifies the action to
    take. Used by startup probe. An example of this argument would be
    ["cat", "/tmp/healthy"]
  container_startup_probe_period_seconds (int): How often (in seconds) to
    perform the startup probe. Default to 10 seconds. Minimum value is 1.
  container_startup_probe_timeout_seconds (int): Number of seconds after
    which the startup probe times out. Defaults to 1 second. Minimum value
    is 1.
  container_health_probe_exec (Sequence[str]): Exec specifies the action to
    take. Used by health probe. An example of this argument would be ["cat",
    "/tmp/healthy"]
  container_health_probe_period_seconds (int): How often (in seconds) to
    perform the health probe. Default to 10 seconds. Minimum value is 1.
  container_health_probe_timeout_seconds (int): Number of seconds after
    which the health probe times out. Defaults to 1 second. Minimum value is
    1.
  explanation_spec: The default explanation specification for this Model.
    The Model can be used for requesting explanation after being deployed if
    it is populated. The Model can be used for batch explanation if it is
    populated. All fields of the explanation_spec can be overridden by
    explanation_spec of DeployModelRequest.deployed_model, or
    explanation_spec of BatchPredictionJob. If the default explanation
    specification is not set for this Model, this Model can still be used
    for requesting explanation by setting explanation_spec of
    DeployModelRequest.deployed_model and for batch explanation by setting
    explanation_spec of BatchPredictionJob.
  parent_model: The resource name of the model into which to upload the
    version. Only specify this field when uploading a new version.
  model_id: The ID to use for the uploaded Model, which will become the
    final component of the model resource name. This value may be up to 63
    characters, and valid characters are `[a-z0-9_-]`. The first character
    cannot be a number or hyphen..
  version_aliases: User provided version aliases so that a model version can
    be referenced via alias (i.e. projects/{project}/locations/{location}/mo
    dels/{model_id}@{version_alias} instead of auto-generated version id
    (i.e.
    projects/{project}/locations/{location}/models/{model_id}@{version_id}).
    The format is a-z{0,126}[a-z0-9] to distinguish from version_id. A
    default version alias will be created for the first version of the
    model, and there must be exactly one default version alias for a model.
  labels: The labels with user-defined metadata to organize your Models.
    Label keys and values can be no longer than 64 characters (Unicode
    codepoints), can only contain lowercase letters, numeric characters,
    underscores and dashes. International characters are allowed. See
    https://goo.gl/xmQnxf for more information and examples of labels.
  base_model_source: A GoogleCloudAiplatformV1beta1ModelBaseModelSource
    object that indicates the source of the model. Currently it only
    supports specifying the Model Garden models and Generative AI Studio
    models.

Returns:
  Response from calling upload model with given request arguments.
healthRouteimageUripredictRoutenamevaluecontainerPortsNcommandexec_periodSecondstimeoutSeconds)artifactUricontainerSpecdescriptionversionDescriptiondisplayNameexplanationSpecbaseModelSourcekeyr"   additionalPropertiesmodelparentModelmodelId)parent.googleCloudAiplatformV1beta1UploadModelRequest)r   .GoogleCloudAiplatformV1beta1ModelContainerSpecr'   args"GoogleCloudAiplatformV1beta1EnvVarenv GoogleCloudAiplatformV1beta1Portports	grpcPortsstrdeploymentTimeoutsharedMemorySizeMb+GoogleCloudAiplatformV1beta1ProbeExecAction!GoogleCloudAiplatformV1beta1ProbestartupProbehealthProbe!GoogleCloudAiplatformV1beta1ModelversionAliasessorteditemsappendLabelsValueAdditionalPropertylabelsr   Upload.AiplatformProjectsLocationsModelsUploadRequestRelativeName.GoogleCloudAiplatformV1beta1UploadModelRequest)%r   
region_refdisplay_namer.   version_descriptionartifact_uricontainer_image_uricontainer_commandcontainer_argscontainer_env_varscontainer_portscontainer_grpc_portscontainer_predict_routecontainer_health_route$container_deployment_timeout_secondscontainer_shared_memory_size_mbcontainer_startup_probe_exec&container_startup_probe_period_seconds'container_startup_probe_timeout_secondscontainer_health_probe_exec%container_health_probe_period_seconds&container_health_probe_timeout_secondsexplanation_specparent_modelmodel_idversion_aliasesrR   base_model_sourcecontainer_speckportstartup_probe_exechealth_probe_execr8   additional_propertiesr4   r"   s%                                        r   UploadV1Beta1ModelsClient.UploadV1Beta1+   s   r 	DD.(0 	E 	
  0n*n =O<Nq --
:
:.q1 ; 3<Nn  &%d --
8
8t
8
L%n  +"*d --
8
8t
8
L*"n ,
2
3c
9 & '*In'$12	%MMEE4 F  	 --
9
9&BD :  ! 	$01	$MMEE3 F  	 --
9
9%AC :    MM;; $. () < E ,e v||~.*#u$$U%6%6%8%K%K5 &L &" 	# / &&4 ' 6el ==DD**,;?==;;(  < " 	E 	#$ $]

"s   'J(=#J-3#J2c                 l   U R                   R                  UUUS9nU(       a  UUl        U(       a  UUl        U	(       a3  U	 Vs/ s H   nU R                   R	                  UU	U   S9PM"     snUl        U
(       a/  U
 Vs/ s H  nU R                   R                  US9PM     snUl        U(       a/  U Vs/ s H  nU R                   R                  US9PM     snUl        U(       a  [        U5      S-   Ul
        U(       a  UUl        U(       d  U(       d  U(       aB  SnU(       a  U R                   R                  US9nU R                   R                  UUUS9Ul        U(       d  U(       d  U(       aB  SnU(       a  U R                   R                  US9nU R                   R                  UUUS9Ul        U R                   R!                  UUUUUUS9n U(       a  UU l        U(       af  / n![%        UR'                  5       5       H3  u  n"n#U!R)                  U R+                  5       R-                  U"U#S	95        M5     U R+                  U!S
9U l        U R0                  R3                  U R                   R5                  UR7                  5       U R                   R9                  U UUS9S95      $ s  snf s  snf s  snf )a#7  Constructs, sends an UploadModel request and returns the LRO to be done.

Args:
  region_ref: The resource reference for a given region. None if the region
    reference is not provided.
  display_name: The display name of the Model. The name can be up to 128
    characters long and can be consist of any UTF-8 characters.
  description: The description of the Model.
  version_description: The description of the Model version.
  artifact_uri: The path to the directory containing the Model artifact and
    any of its supporting files. Not present for AutoML Models.
  container_image_uri: Immutable. URI of the Docker image to be used as the
    custom container for serving predictions. This URI must identify an
    image in Artifact Registry or Container Registry. Learn more about the
    [container publishing requirements](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#publishing), including
    permissions requirements for the Vertex AI Service Agent. The container
    image is ingested upon ModelService.UploadModel, stored internally, and
    this original path is afterwards not used. To learn about the
    requirements for the Docker image itself, see [Custom container
    requirements](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#). You can use the URI
    to one of Vertex AI's [pre-built container images for
    prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-
    built-containers) in this field.
  container_command: Specifies the command that runs when the container
    starts. This overrides the container's [ENTRYPOINT](https://docs.docker.
    com/engine/reference/builder/#entrypoint). Specify this field as an
    array of executable and arguments, similar to a Docker `ENTRYPOINT`'s
    "exec" form, not its "shell" form. If you do not specify this field,
    then the container's `ENTRYPOINT` runs, in conjunction with the args
    field or the container's
    [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd), if
    either exists. If this field is not specified and the container does not
    have an `ENTRYPOINT`, then refer to the Docker documentation about [how
    `CMD` and `ENTRYPOINT`
    interact](https://docs.docker.com/engine/reference/builder/#understand-
    how-cmd-and-entrypoint-interact). If you specify this field, then you
    can also specify the `args` field to provide additional arguments for
    this command. However, if you specify this field, then the container's
    `CMD` is ignored. See the [Kubernetes documentation about how the
    `command` and `args` fields interact with a container's `ENTRYPOINT` and
    `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-
    command-argument-container/#notes). In this field, you can reference
    [environment variables set by Vertex
    AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables) and environment variables set in
    the env field. You cannot reference environment variables set in the
    Docker image. In order for environment variables to be expanded,
    reference them by using the following syntax: $( VARIABLE_NAME) Note
    that this differs from Bash variable expansion, which does not use
    parentheses. If a variable cannot be resolved, the reference in the
    input string is used unchanged. To avoid variable expansion, you can
    escape this syntax with `$$`; for example: $$(VARIABLE_NAME) This field
    corresponds to the `command` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_args: Specifies arguments for the command that runs when the
    container starts. This overrides the container's
    [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd). Specify
    this field as an array of executable and arguments, similar to a Docker
    `CMD`'s "default parameters" form. If you don't specify this field but
    do specify the command field, then the command from the `command` field
    runs without any additional arguments. See the [Kubernetes documentation
    about how the `command` and `args` fields interact with a container's
    `ENTRYPOINT` and `CMD`](https://kubernetes.io/docs/tasks/inject-data-
    application/define-command-argument-container/#notes). If you don't
    specify this field and don't specify the `command` field, then the
    container's
    [`ENTRYPOINT`](https://docs.docker.com/engine/reference/builder/#cmd)
    and `CMD` determine what runs based on their default behavior. See the
    Docker documentation about [how `CMD` and `ENTRYPOINT`
    interact](https://docs.docker.com/engine/reference/builder/#understand-
    how-cmd-and-entrypoint-interact). In this field, you can reference
    [environment variables set by Vertex
    AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables) and environment variables set in
    the env field. You cannot reference environment variables set in the
    Docker image. In order for environment variables to be expanded,
    reference them by using the following syntax: $( VARIABLE_NAME) Note
    that this differs from Bash variable expansion, which does not use
    parentheses. If a variable cannot be resolved, the reference in the
    input string is used unchanged. To avoid variable expansion, you can
    escape this syntax with `$$`; for example: $$(VARIABLE_NAME) This field
    corresponds to the `args` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core)..
  container_env_vars: List of environment variables to set in the container.
    After the container starts running, code running in the container can
    read these environment variables. Additionally, the command and args
    fields can reference these variables. Later entries in this list can
    also reference earlier entries. For example, the following example sets
    the variable `VAR_2` to have the value `foo bar`: ```json [ { "name":
    "VAR_1", "value": "foo" }, { "name": "VAR_2", "value": "$(VAR_1) bar" }
    ] ``` If you switch the order of the variables in the example, then the
    expansion does not occur. This field corresponds to the `env` field of
    the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_ports: List of ports to expose from the container. Vertex AI
    sends any http prediction requests that it receives to the first port on
    this list. Vertex AI also sends [liveness and health
    checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#liveness) to this port. If you do not specify
    this field, it defaults to following value: ```json [ { "containerPort":
    8080 } ] ``` Vertex AI does not use ports other than the first one
    listed. This field corresponds to the `ports` field of the Kubernetes
    Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_grpc_ports: List of ports to expose from the container. Vertex
    AI sends any grpc prediction requests that it receives to the first port
    on this list. Vertex AI also sends [liveness and health
    checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#liveness) to this port. If you do not specify
    this field, gRPC requests to the container will be disabled. Vertex AI
    does not use ports other than the first one listed. This field
    corresponds to the `ports` field of the Kubernetes Containers [v1 core
    API](https://kubernetes.io/docs/reference/generated/kubernetes-
    api/v1.23/#container-v1-core).
  container_predict_route: HTTP path on the container to send prediction
    requests to. Vertex AI forwards requests sent using
    projects.locations.endpoints.predict to this path on the container's IP
    address and port. Vertex AI then returns the container's response in the
    API response. For example, if you set this field to `/foo`, then when
    Vertex AI receives a prediction request, it forwards the request body in
    a POST request to the `/foo` path on the port of your container
    specified by the first value of this `ModelContainerSpec`'s ports field.
    If you don't specify this field, it defaults to the following value when
    you deploy this Model to an Endpoint:
    /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The
    placeholders in this value are replaced as follows: * ENDPOINT: The last
    segment (following `endpoints/`)of the Endpoint.name][] field of the
    Endpoint where this Model has been deployed. (Vertex AI makes this value
    available to your container code as the [`AIP_ENDPOINT_ID` environment
    variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables).) * DEPLOYED_MODEL:
    DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value
    available to your container code as the [`AIP_DEPLOYED_MODEL_ID`
    environment variable](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#aip-variables).)
  container_health_route: HTTP path on the container to send health checks
    to. Vertex AI intermittently sends GET requests to this path on the
    container's IP address and port to check that the container is healthy.
    Read more about [health checks](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#health). For example,
    if you set this field to `/bar`, then Vertex AI intermittently sends a
    GET request to the `/bar` path on the port of your container specified
    by the first value of this `ModelContainerSpec`'s ports field. If you
    don't specify this field, it defaults to the following value when you
    deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/
    DEPLOYED_MODEL:predict The placeholders in this value are replaced as
    follows * ENDPOINT: The last segment (following `endpoints/`)of the
    Endpoint.name][] field of the Endpoint where this Model has been
    deployed. (Vertex AI makes this value available to your container code
    as the [`AIP_ENDPOINT_ID` environment
    variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-
    container-requirements#aip-variables).) * DEPLOYED_MODEL:
    DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value
    available to your container code as the [`AIP_DEPLOYED_MODEL_ID`
    environment variable](https://cloud.google.com/vertex-
    ai/docs/predictions/custom-container-requirements#aip-variables).)
  container_deployment_timeout_seconds (int): Deployment timeout in seconds.
  container_shared_memory_size_mb (int): The amount of the VM memory to
    reserve as the shared memory for the model in megabytes.
  container_startup_probe_exec (Sequence[str]): Exec specifies the action to
    take. Used by startup probe. An example of this argument would be
    ["cat", "/tmp/healthy"]
  container_startup_probe_period_seconds (int): How often (in seconds) to
    perform the startup probe. Default to 10 seconds. Minimum value is 1.
  container_startup_probe_timeout_seconds (int): Number of seconds after
    which the startup probe times out. Defaults to 1 second. Minimum value
    is 1.
  container_health_probe_exec (Sequence[str]): Exec specifies the action to
    take. Used by health probe. An example of this argument would be ["cat",
    "/tmp/healthy"]
  container_health_probe_period_seconds (int): How often (in seconds) to
    perform the health probe. Default to 10 seconds. Minimum value is 1.
  container_health_probe_timeout_seconds (int): Number of seconds after
    which the health probe times out. Defaults to 1 second. Minimum value is
    1.
  explanation_spec: The default explanation specification for this Model.
    The Model can be used for requesting explanation after being deployed if
    it is populated. The Model can be used for batch explanation if it is
    populated. All fields of the explanation_spec can be overridden by
    explanation_spec of DeployModelRequest.deployed_model, or
    explanation_spec of BatchPredictionJob. If the default explanation
    specification is not set for this Model, this Model can still be used
    for requesting explanation by setting explanation_spec of
    DeployModelRequest.deployed_model and for batch explanation by setting
    explanation_spec of BatchPredictionJob.
  parent_model: The resource name of the model into which to upload the
    version. Only specify this field when uploading a new version.
  model_id: The ID to use for the uploaded Model, which will become the
    final component of the model resource name. This value may be up to 63
    characters, and valid characters are `[a-z0-9_-]`. The first character
    cannot be a number or hyphen..
  version_aliases: User provided version aliases so that a model version can
    be referenced via alias (i.e. projects/{project}/locations/{location}/mo
    dels/{model_id}@{version_alias} instead of auto-generated version id
    (i.e.
    projects/{project}/locations/{location}/models/{model_id}@{version_id}).
    The format is a-z{0,126}[a-z0-9] to distinguish from version_id. A
    default version alias will be created for the first version of the
    model, and there must be exactly one default version alias for a model.
  labels: The labels with user-defined metadata to organize your Models.
    Label keys and values can be no longer than 64 characters (Unicode
    codepoints), can only contain lowercase letters, numeric characters,
    underscores and dashes. International characters are allowed. See
    https://goo.gl/xmQnxf for more information and examples of labels.

Returns:
  Response from calling upload model with given request arguments.
r   r    r#   r%   Nr&   r(   )r,   r-   r.   r/   r0   r1   r3   r5   r7   )r;   )googleCloudAiplatformV1UploadModelRequest)r   )GoogleCloudAiplatformV1ModelContainerSpecr'   r>   GoogleCloudAiplatformV1EnvVarr@   GoogleCloudAiplatformV1PortrB   rC   rD   rE   rF   &GoogleCloudAiplatformV1ProbeExecActionGoogleCloudAiplatformV1ProberI   rJ   GoogleCloudAiplatformV1ModelrL   rM   rN   rO   rP   rQ   rR   r   rS   rT   rU   )GoogleCloudAiplatformV1UploadModelRequest)$r   rW   rX   r.   rY   rZ   r[   r\   r]   r^   r_   r`   ra   rb   rc   rd   re   rf   rg   rh   ri   rj   rk   rl   rm   rn   rR   rp   rq   rr   rs   rt   r8   ru   r4   r"   s$                                       r   UploadV1ModelsClient.UploadV1  s   b ]]LL*$, M .N 0n*n =O<Nq --
5
5.q1 6 3<Nn  &%d --
3
3$
3
G%n  +"*d --
3
3$
3
G*"n ,
2
3c
9 & '*In'$12	%MM@@4 A  	 --
4
4&BD 5  ! 	$01	$MM@@3 A  	 --
4
4%AC 5    MM66 $. ( 7 *E ,e v||~.*#u$$U%6%6%8%K%K5 &L &" 	# / &&4 ' 6el ==DD**,6:mm66(  7 " 	E 	#$ $Y

"s   'J'=#J,3#J1c                     U R                   R                  UR                  5       S9nU R                  R	                  U5      $ )zGets (describe) the given model.

Args:
  model_ref: The resource reference for a given model. None if model
    resource reference is not provided.

Returns:
  Response from calling get model with request containing given model.
r!   )r   +AiplatformProjectsLocationsModelsGetRequestrU   r   Getr   	model_refrequests      r   r   ModelsClient.Get  s?     mmGG##% H 'G==W%%r   c                     U R                   R                  UR                  5       S9nU R                  R	                  U5      $ )zDeletes the given model.

Args:
  model_ref: The resource reference for a given model. None if model
    resource reference is not provided.

Returns:
  Response from calling delete model with request containing given model.
r   )r   .AiplatformProjectsLocationsModelsDeleteRequestrU   r   Deleter   s      r   r   ModelsClient.Delete  s?     mmJJ##% K 'G==((r   c                     U R                   R                  UR                  5       S9nU R                  R	                  U5      $ )zDeletes the given model version.

Args:
  model_version_ref: The resource reference for a given model version.

Returns:
  Response from calling delete version with request containing given model
  version.
r   )r   5AiplatformProjectsLocationsModelsDeleteVersionRequestrU   r   DeleteVersion)r   model_version_refr   s      r   r   ModelsClient.DeleteVersion  sD     	KK"//1 	L 	
 
 ==&&w//r   c                     [         R                  " U R                  U R                  R	                  UR                  5       S9SSUS9$ )ab  List all models in the given region.

Args:
  limit: int, The maximum number of records to yield. None if all available
    records should be yielded.
  region_ref: The resource reference for a given region. None if the region
    reference is not provided.

Returns:
  Response from calling list models with request containing given models
  and limit.
)r;   modelspageSize)fieldbatch_size_attributelimit)r   YieldFromListr   r   ,AiplatformProjectsLocationsModelsListRequestrU   )r   r   rW   s      r   ListModelsClient.List  sK     ##BB**, 	C 	.' r   c           	          [         R                  " U R                  U R                  R	                  UR                  5       S9SSSUS9$ )ar  List all model versions of the given model.

Args:
  model_ref: The resource reference for a given model. None if model
    resource reference is not provided.
  limit: int, The maximum number of records to yield. None if all available
    records should be yielded.

Returns:
  Response from calling list model versions with request containing given
  model and limit.
r   ListVersionsr   r   )methodr   r   r   )r   r   r   r   4AiplatformProjectsLocationsModelsListVersionsRequestrU   )r   r   r   s      r   ListVersionModelsClient.ListVersion  sN     ##JJ'') 	K 	+' r   c           
          SnU(       a  U R                   R                  US9nU R                   R                  UR                  5       U R                   R	                  UUUUS9S9nU R
                  R                  U5      $ )a  Copies the given source model into specified location.

The source model is copied into specified location (including cross-region)
either as a new model or a new model version under given parent model.

Args:
  destination_region_ref: the resource reference to the location into which
    to copy the Model.
  source_model: The resource name of the Model to copy.
  kms_key_name: The KMS key name for specifying encryption spec.
  destination_model_id: The destination model resource name to copy the
    model into.
  destination_parent_model: The destination parent model to copy the model
    as a model version into.

Returns:
  Response from calling copy model.
N
kmsKeyNamesourceModelencryptionSpecr9   r:   )r;   ,googleCloudAiplatformV1beta1CopyModelRequest)r   *GoogleCloudAiplatformV1beta1EncryptionSpec,AiplatformProjectsLocationsModelsCopyRequestrU   ,GoogleCloudAiplatformV1beta1CopyModelRequestr   Copyr   destination_region_refsource_modelkms_key_namedestination_model_iddestination_parent_modelencryption_specr   s           r   CopyV1Beta1ModelsClient.CopyV1Beta1-  s    0 O
--
B
B% C  
 mmHH%22459]]	5	5$*0(	 
6 
* I +G ==g&&r   c           
          SnU(       a  U R                   R                  US9nU R                   R                  UR                  5       U R                   R	                  UUUUS9S9nU R
                  R                  U5      $ )a8  Copies the given source model into specified location.

The source model is copied into specified location (including cross-region)
either as a new model or a new model version under given parent model.

Args:
  destination_region_ref: the resource reference to the location into which
    to copy the Model.
  source_model: The resource name of the Model to copy.
  kms_key_name: The name of the KMS key to use for model encryption.
  destination_model_id: Optional. Thew custom ID to be used as the resource
    name of the new model. This value may be up to 63 characters, and valid
    characters are  `[a-z0-9_-]`. The first character cannot be a number or
    hyphen.
  destination_parent_model: The destination parent model to copy the model
    as a model version into.

Returns:
  Response from calling copy model.
Nr   r   )r;   'googleCloudAiplatformV1CopyModelRequest)r   %GoogleCloudAiplatformV1EncryptionSpecr   rU   'GoogleCloudAiplatformV1CopyModelRequestr   r   r   s           r   CopyV1ModelsClient.CopyV1V  s    4 O
--
=
=% >  
 mmHH%22404	0	0$*0(	 
1 
* I +G ==g&&r   )r   r   r   )NN)NNNNNNNNNNNNNNNNNNNNNNNNNNN)NNNNNNNNNNNNNNNNNNNNNNNNNN)NNNNN)__name__
__module____qualname____firstlineno____doc__r   rv   r   r   r   r   r   r   r   r   __static_attributes__ r   r   r
   r
      s   : "!+/&*#'-1.2"&,0-19Y$x
  #' #'!%""&#$('+&*48/3,06:7;+/596: $ #5M$^
&)0"*. *.##'++/''T %)"&&*)'r   r
   N)r   
__future__r   r   r   apitools.base.pyr   googlecloudsdk.api_lib.utilr   googlecloudsdk.command_lib.air   objectr
   r   r   r   <module>r      s*    , &  ' ' , 3e'6 e'r   