Get desktop application:
View/edit binary Protocol Buffers messages
The service that manages Vertex AI Dataset and its child resources.
Creates a Dataset.
Request message for [DatasetService.CreateDataset][google.cloud.aiplatform.v1.DatasetService.CreateDataset].
Required. The resource name of the Location to create the Dataset in. Format: `projects/{project}/locations/{location}`
Required. The Dataset to create.
Gets a Dataset.
Request message for [DatasetService.GetDataset][google.cloud.aiplatform.v1.DatasetService.GetDataset]. Next ID: 4
Required. The name of the Dataset resource.
Mask specifying which fields to read.
Updates a Dataset.
Request message for [DatasetService.UpdateDataset][google.cloud.aiplatform.v1.DatasetService.UpdateDataset].
Required. The Dataset which replaces the resource on the server.
Required. The update mask applies to the resource. For the `FieldMask` definition, see [google.protobuf.FieldMask][google.protobuf.FieldMask]. Updatable fields: * `display_name` * `description` * `labels`
Lists Datasets in a Location.
Request message for [DatasetService.ListDatasets][google.cloud.aiplatform.v1.DatasetService.ListDatasets].
Required. The name of the Dataset's parent resource. Format: `projects/{project}/locations/{location}`
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `display_name`: supports = and != * `metadata_schema_uri`: supports = and != * `labels` supports general map functions that is: * `labels.key=value` - key:value equality * `labels.key:* or labels:key - key existence * A key including a space must be quoted. `labels."a key"`. Some examples: * `displayName="myDisplayName"` * `labels.myKey="myValue"`
The standard list page size.
The standard list page token.
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time`
Response message for [DatasetService.ListDatasets][google.cloud.aiplatform.v1.DatasetService.ListDatasets].
A list of Datasets that matches the specified filter in the request.
The standard List next-page token.
Deletes a Dataset.
Request message for [DatasetService.DeleteDataset][google.cloud.aiplatform.v1.DatasetService.DeleteDataset].
Required. The resource name of the Dataset to delete. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Imports data into a Dataset.
Request message for [DatasetService.ImportData][google.cloud.aiplatform.v1.DatasetService.ImportData].
Required. The name of the Dataset resource. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Required. The desired input locations. The contents of all input locations will be imported in one batch.
Exports data from a Dataset.
Request message for [DatasetService.ExportData][google.cloud.aiplatform.v1.DatasetService.ExportData].
Required. The name of the Dataset resource. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Required. The desired output location.
Create a version from a Dataset.
Request message for [DatasetService.CreateDatasetVersion][google.cloud.aiplatform.v1.DatasetService.CreateDatasetVersion].
Required. The name of the Dataset resource. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Required. The version to be created. The same CMEK policies with the original Dataset will be applied the dataset version. So here we don't need to specify the EncryptionSpecType here.
Updates a DatasetVersion.
Request message for [DatasetService.UpdateDatasetVersion][google.cloud.aiplatform.v1.DatasetService.UpdateDatasetVersion].
Required. The DatasetVersion which replaces the resource on the server.
Required. The update mask applies to the resource. For the `FieldMask` definition, see [google.protobuf.FieldMask][google.protobuf.FieldMask]. Updatable fields: * `display_name`
Deletes a Dataset version.
Request message for [DatasetService.DeleteDatasetVersion][google.cloud.aiplatform.v1.DatasetService.DeleteDatasetVersion].
Required. The resource name of the Dataset version to delete. Format: `projects/{project}/locations/{location}/datasets/{dataset}/datasetVersions/{dataset_version}`
Gets a Dataset version.
Request message for [DatasetService.GetDatasetVersion][google.cloud.aiplatform.v1.DatasetService.GetDatasetVersion]. Next ID: 4
Required. The resource name of the Dataset version to delete. Format: `projects/{project}/locations/{location}/datasets/{dataset}/datasetVersions/{dataset_version}`
Mask specifying which fields to read.
Lists DatasetVersions in a Dataset.
Request message for [DatasetService.ListDatasetVersions][google.cloud.aiplatform.v1.DatasetService.ListDatasetVersions].
Required. The resource name of the Dataset to list DatasetVersions from. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Optional. The standard list filter.
Optional. The standard list page size.
Optional. The standard list page token.
Optional. Mask specifying which fields to read.
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
Response message for [DatasetService.ListDatasetVersions][google.cloud.aiplatform.v1.DatasetService.ListDatasetVersions].
A list of DatasetVersions that matches the specified filter in the request.
The standard List next-page token.
Restores a dataset version.
Request message for [DatasetService.RestoreDatasetVersion][google.cloud.aiplatform.v1.DatasetService.RestoreDatasetVersion].
Required. The name of the DatasetVersion resource. Format: `projects/{project}/locations/{location}/datasets/{dataset}/datasetVersions/{dataset_version}`
Lists DataItems in a Dataset.
Request message for [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems].
Required. The resource name of the Dataset to list DataItems from. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
The standard list filter.
The standard list page size.
The standard list page token.
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
Response message for [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems].
A list of DataItems that matches the specified filter in the request.
The standard List next-page token.
Searches DataItems in a Dataset.
Request message for [DatasetService.SearchDataItems][google.cloud.aiplatform.v1.DatasetService.SearchDataItems].
A comma-separated list of data item fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
Expression that allows ranking results based on annotation's property.
Required. The resource name of the Dataset from which to search DataItems. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
The resource name of a SavedQuery(annotation set in UI). Format: `projects/{project}/locations/{location}/datasets/{dataset}/savedQueries/{saved_query}` All of the search will be done in the context of this SavedQuery.
The resource name of a DataLabelingJob. Format: `projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}` If this field is set, all of the search will be done in the context of this DataLabelingJob.
An expression for filtering the DataItem that will be returned. * `data_item_id` - for = or !=. * `labeled` - for = or !=. * `has_annotation(ANNOTATION_SPEC_ID)` - true only for DataItem that have at least one annotation with annotation_spec_id = `ANNOTATION_SPEC_ID` in the context of SavedQuery or DataLabelingJob. For example: * `data_item=1` * `has_annotation(5)`
An expression for filtering the Annotations that will be returned per DataItem. * `annotation_spec_id` - for = or !=.
An expression that specifies what Annotations will be returned per DataItem. Annotations satisfied either of the conditions will be returned. * `annotation_spec_id` - for = or !=. Must specify `saved_query_id=` - saved query id that annotations should belong to.
Mask specifying which fields of [DataItemView][google.cloud.aiplatform.v1.DataItemView] to read.
If set, only up to this many of Annotations will be returned per DataItemView. The maximum value is 1000. If not set, the maximum value will be used.
Requested page size. Server may return fewer results than requested. Default and maximum page size is 100.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
A token identifying a page of results for the server to return Typically obtained via [SearchDataItemsResponse.next_page_token][google.cloud.aiplatform.v1.SearchDataItemsResponse.next_page_token] of the previous [DatasetService.SearchDataItems][google.cloud.aiplatform.v1.DatasetService.SearchDataItems] call.
Response message for [DatasetService.SearchDataItems][google.cloud.aiplatform.v1.DatasetService.SearchDataItems].
The DataItemViews read.
A token to retrieve next page of results. Pass to [SearchDataItemsRequest.page_token][google.cloud.aiplatform.v1.SearchDataItemsRequest.page_token] to obtain that page.
Lists SavedQueries in a Dataset.
Request message for [DatasetService.ListSavedQueries][google.cloud.aiplatform.v1.DatasetService.ListSavedQueries].
Required. The resource name of the Dataset to list SavedQueries from. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
The standard list filter.
The standard list page size.
The standard list page token.
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
Response message for [DatasetService.ListSavedQueries][google.cloud.aiplatform.v1.DatasetService.ListSavedQueries].
A list of SavedQueries that match the specified filter in the request.
The standard List next-page token.
Deletes a SavedQuery.
Request message for [DatasetService.DeleteSavedQuery][google.cloud.aiplatform.v1.DatasetService.DeleteSavedQuery].
Required. The resource name of the SavedQuery to delete. Format: `projects/{project}/locations/{location}/datasets/{dataset}/savedQueries/{saved_query}`
Gets an AnnotationSpec.
Request message for [DatasetService.GetAnnotationSpec][google.cloud.aiplatform.v1.DatasetService.GetAnnotationSpec].
Required. The name of the AnnotationSpec resource. Format: `projects/{project}/locations/{location}/datasets/{dataset}/annotationSpecs/{annotation_spec}`
Mask specifying which fields to read.
Identifies a concept with which DataItems may be annotated with.
Output only. Resource name of the AnnotationSpec.
Required. The user-defined name of the AnnotationSpec. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Output only. Timestamp when this AnnotationSpec was created.
Output only. Timestamp when AnnotationSpec was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Lists Annotations belongs to a dataitem This RPC is only available in InternalDatasetService. It is only used for exporting conversation data to CCAI Insights.
Request message for [DatasetService.ListAnnotations][google.cloud.aiplatform.v1.DatasetService.ListAnnotations].
Required. The resource name of the DataItem to list Annotations from. Format: `projects/{project}/locations/{location}/datasets/{dataset}/dataItems/{data_item}`
The standard list filter.
The standard list page size.
The standard list page token.
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending.
Response message for [DatasetService.ListAnnotations][google.cloud.aiplatform.v1.DatasetService.ListAnnotations].
A list of Annotations that matches the specified filter in the request.
The standard List next-page token.
A service that manages the DeploymentResourcePool resource.
Create a DeploymentResourcePool.
Request message for CreateDeploymentResourcePool method.
Required. The parent location resource where this DeploymentResourcePool will be created. Format: `projects/{project}/locations/{location}`
Required. The DeploymentResourcePool to create.
Required. The ID to use for the DeploymentResourcePool, which will become the final component of the DeploymentResourcePool's resource name. The maximum length is 63 characters, and valid characters are `/^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$/`.
Get a DeploymentResourcePool.
Request message for GetDeploymentResourcePool method.
Required. The name of the DeploymentResourcePool to retrieve. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
List DeploymentResourcePools in a location.
Request message for ListDeploymentResourcePools method.
Required. The parent Location which owns this collection of DeploymentResourcePools. Format: `projects/{project}/locations/{location}`
The maximum number of DeploymentResourcePools to return. The service may return fewer than this value.
A page token, received from a previous `ListDeploymentResourcePools` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListDeploymentResourcePools` must match the call that provided the page token.
Response message for ListDeploymentResourcePools method.
The DeploymentResourcePools from the specified location.
A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
Update a DeploymentResourcePool.
Request message for UpdateDeploymentResourcePool method.
Required. The DeploymentResourcePool to update. The DeploymentResourcePool's `name` field is used to identify the DeploymentResourcePool to update. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
Required. The list of fields to update.
Delete a DeploymentResourcePool.
Request message for DeleteDeploymentResourcePool method.
Required. The name of the DeploymentResourcePool to delete. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
List DeployedModels that have been deployed on this DeploymentResourcePool.
Request message for QueryDeployedModels method.
Required. The name of the target DeploymentResourcePool to query. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
The maximum number of DeployedModels to return. The service may return fewer than this value.
A page token, received from a previous `QueryDeployedModels` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `QueryDeployedModels` must match the call that provided the page token.
Response message for QueryDeployedModels method.
DEPRECATED Use deployed_model_refs instead.
A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
References to the DeployedModels that share the specified deploymentResourcePool.
The total number of DeployedModels on this DeploymentResourcePool.
The total number of Endpoints that have DeployedModels on this DeploymentResourcePool.
A service for managing Vertex AI's Endpoints.
Creates an Endpoint.
Request message for [EndpointService.CreateEndpoint][google.cloud.aiplatform.v1.EndpointService.CreateEndpoint].
Required. The resource name of the Location to create the Endpoint in. Format: `projects/{project}/locations/{location}`
Required. The Endpoint to create.
Immutable. The ID to use for endpoint, which will become the final component of the endpoint resource name. If not provided, Vertex AI will generate a value for this ID. If the first character is a letter, this value may be up to 63 characters, and valid characters are `[a-z0-9-]`. The last character must be a letter or number. If the first character is a number, this value may be up to 9 characters, and valid characters are `[0-9]` with no leading zeros. When using HTTP/JSON, this field is populated based on a query string argument, such as `?endpoint_id=12345`. This is the fallback for fields that are not included in either the URI or the body.
Gets an Endpoint.
Request message for [EndpointService.GetEndpoint][google.cloud.aiplatform.v1.EndpointService.GetEndpoint]
Required. The name of the Endpoint resource. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Lists Endpoints in a Location.
Request message for [EndpointService.ListEndpoints][google.cloud.aiplatform.v1.EndpointService.ListEndpoints].
Required. The resource name of the Location from which to list the Endpoints. Format: `projects/{project}/locations/{location}`
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `endpoint` supports `=` and `!=`. `endpoint` represents the Endpoint ID, i.e. the last segment of the Endpoint's [resource name][google.cloud.aiplatform.v1.Endpoint.name]. * `display_name` supports `=` and `!=`. * `labels` supports general map functions that is: * `labels.key=value` - key:value equality * `labels.key:*` or `labels:key` - key existence * A key including a space must be quoted. `labels."a key"`. * `base_model_name` only supports `=`. Some examples: * `endpoint=1` * `displayName="myDisplayName"` * `labels.myKey="myValue"` * `baseModelName="text-bison"`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListEndpointsResponse.next_page_token][google.cloud.aiplatform.v1.ListEndpointsResponse.next_page_token] of the previous [EndpointService.ListEndpoints][google.cloud.aiplatform.v1.EndpointService.ListEndpoints] call.
Optional. Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time` Example: `display_name, create_time desc`.
Response message for [EndpointService.ListEndpoints][google.cloud.aiplatform.v1.EndpointService.ListEndpoints].
List of Endpoints in the requested page.
A token to retrieve the next page of results. Pass to [ListEndpointsRequest.page_token][google.cloud.aiplatform.v1.ListEndpointsRequest.page_token] to obtain that page.
Updates an Endpoint.
Request message for [EndpointService.UpdateEndpoint][google.cloud.aiplatform.v1.EndpointService.UpdateEndpoint].
Required. The Endpoint which replaces the resource on the server.
Required. The update mask applies to the resource. See [google.protobuf.FieldMask][google.protobuf.FieldMask].
Updates an Endpoint with a long running operation.
Request message for [EndpointService.UpdateEndpointLongRunning][google.cloud.aiplatform.v1.EndpointService.UpdateEndpointLongRunning].
Required. The Endpoint which replaces the resource on the server. Currently we only support updating the `client_connection_config` field, all the other fields' update will be blocked.
Deletes an Endpoint.
Request message for [EndpointService.DeleteEndpoint][google.cloud.aiplatform.v1.EndpointService.DeleteEndpoint].
Required. The name of the Endpoint resource to be deleted. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Deploys a Model into this Endpoint, creating a DeployedModel within it.
Request message for [EndpointService.DeployModel][google.cloud.aiplatform.v1.EndpointService.DeployModel].
Required. The name of the Endpoint resource into which to deploy a Model. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The DeployedModel to be created within the Endpoint. Note that [Endpoint.traffic_split][google.cloud.aiplatform.v1.Endpoint.traffic_split] must be updated for the DeployedModel to start receiving traffic, either as part of this call, or via [EndpointService.UpdateEndpoint][google.cloud.aiplatform.v1.EndpointService.UpdateEndpoint].
A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If this field is non-empty, then the Endpoint's [traffic_split][google.cloud.aiplatform.v1.Endpoint.traffic_split] will be overwritten with it. To refer to the ID of the just being deployed Model, a "0" should be used, and the actual ID of the new DeployedModel will be filled in its place by this method. The traffic percentage values must add up to 100. If this field is empty, then the Endpoint's [traffic_split][google.cloud.aiplatform.v1.Endpoint.traffic_split] is not updated.
Undeploys a Model from an Endpoint, removing a DeployedModel from it, and freeing all resources it's using.
Request message for [EndpointService.UndeployModel][google.cloud.aiplatform.v1.EndpointService.UndeployModel].
Required. The name of the Endpoint resource from which to undeploy a Model. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The ID of the DeployedModel to be undeployed from the Endpoint.
If this field is provided, then the Endpoint's [traffic_split][google.cloud.aiplatform.v1.Endpoint.traffic_split] will be overwritten with it. If last DeployedModel is being undeployed from the Endpoint, the [Endpoint.traffic_split] will always end up empty when this call returns. A DeployedModel will be successfully undeployed only if it doesn't have any traffic assigned to it when this method executes, or if this field unassigns any traffic to it.
Updates an existing deployed model. Updatable fields include `min_replica_count`, `max_replica_count`, `autoscaling_metric_specs`, `disable_container_logging` (v1 only), and `enable_container_logging` (v1beta1 only).
Request message for [EndpointService.MutateDeployedModel][google.cloud.aiplatform.v1.EndpointService.MutateDeployedModel].
Required. The name of the Endpoint resource into which to mutate a DeployedModel. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The DeployedModel to be mutated within the Endpoint. Only the following fields can be mutated: * `min_replica_count` in either [DedicatedResources][google.cloud.aiplatform.v1.DedicatedResources] or [AutomaticResources][google.cloud.aiplatform.v1.AutomaticResources] * `max_replica_count` in either [DedicatedResources][google.cloud.aiplatform.v1.DedicatedResources] or [AutomaticResources][google.cloud.aiplatform.v1.AutomaticResources] * [autoscaling_metric_specs][google.cloud.aiplatform.v1.DedicatedResources.autoscaling_metric_specs] * `disable_container_logging` (v1 only) * `enable_container_logging` (v1beta1 only)
Required. The update mask applies to the resource. See [google.protobuf.FieldMask][google.protobuf.FieldMask].
Vertex AI Online Evaluation Service.
Evaluates instances based on a given metric.
Request message for EvaluationService.EvaluateInstances.
Instances and specs for evaluation
Auto metric instances. Instances and metric spec for exact match metric.
Instances and metric spec for bleu metric.
Instances and metric spec for rouge metric.
LLM-based metric instance. General text generation metrics, applicable to other categories. Input for fluency metric.
Input for coherence metric.
Input for safety metric.
Input for groundedness metric.
Input for fulfillment metric.
Input for summarization quality metric.
Input for pairwise summarization quality metric.
Input for summarization helpfulness metric.
Input for summarization verbosity metric.
Input for question answering quality metric.
Input for pairwise question answering quality metric.
Input for question answering relevance metric.
Input for question answering helpfulness metric.
Input for question answering correctness metric.
Input for pointwise metric.
Input for pairwise metric.
Tool call metric instances. Input for tool call valid metric.
Input for tool name match metric.
Input for tool parameter key match metric.
Input for tool parameter key value match metric.
Translation metrics. Input for Comet metric.
Input for Metricx metric.
Required. The resource name of the Location to evaluate the instances. Format: `projects/{project}/locations/{location}`
Response message for EvaluationService.EvaluateInstances.
Evaluation results will be served in the same order as presented in EvaluationRequest.instances.
Auto metric evaluation results. Results for exact match metric.
Results for bleu metric.
Results for rouge metric.
LLM-based metric evaluation result. General text generation metrics, applicable to other categories. Result for fluency metric.
Result for coherence metric.
Result for safety metric.
Result for groundedness metric.
Result for fulfillment metric.
Summarization only metrics. Result for summarization quality metric.
Result for pairwise summarization quality metric.
Result for summarization helpfulness metric.
Result for summarization verbosity metric.
Question answering only metrics. Result for question answering quality metric.
Result for pairwise question answering quality metric.
Result for question answering relevance metric.
Result for question answering helpfulness metric.
Result for question answering correctness metric.
Generic metrics. Result for pointwise metric.
Result for pairwise metric.
Tool call metrics. Results for tool call valid metric.
Results for tool name match metric.
Results for tool parameter key match metric.
Results for tool parameter key value match metric.
Translation metrics. Result for Comet metric.
Result for Metricx metric.
The service that handles CRUD and List for resources for FeatureOnlineStore.
Creates a new FeatureOnlineStore in a given project and location.
Request message for [FeatureOnlineStoreAdminService.CreateFeatureOnlineStore][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.CreateFeatureOnlineStore].
Required. The resource name of the Location to create FeatureOnlineStores. Format: `projects/{project}/locations/{location}`
Required. The FeatureOnlineStore to create.
Required. The ID to use for this FeatureOnlineStore, which will become the final component of the FeatureOnlineStore's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within the project and location.
Gets details of a single FeatureOnlineStore.
Request message for [FeatureOnlineStoreAdminService.GetFeatureOnlineStore][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.GetFeatureOnlineStore].
Required. The name of the FeatureOnlineStore resource.
Lists FeatureOnlineStores in a given project and location.
Request message for [FeatureOnlineStoreAdminService.ListFeatureOnlineStores][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureOnlineStores].
Required. The resource name of the Location to list FeatureOnlineStores. Format: `projects/{project}/locations/{location}`
Lists the FeatureOnlineStores that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality and key presence. Examples: * `create_time > "2020-01-01" OR update_time > "2020-01-01"` FeatureOnlineStores created or updated after 2020-01-01. * `labels.env = "prod"` FeatureOnlineStores with label "env" set to "prod".
The maximum number of FeatureOnlineStores to return. The service may return fewer than this value. If unspecified, at most 100 FeatureOnlineStores will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
A page token, received from a previous [FeatureOnlineStoreAdminService.ListFeatureOnlineStores][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureOnlineStores] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeatureOnlineStoreAdminService.ListFeatureOnlineStores][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureOnlineStores] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields: * `create_time` * `update_time`
Response message for [FeatureOnlineStoreAdminService.ListFeatureOnlineStores][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureOnlineStores].
The FeatureOnlineStores matching the request.
A token, which can be sent as [ListFeatureOnlineStoresRequest.page_token][google.cloud.aiplatform.v1.ListFeatureOnlineStoresRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Updates the parameters of a single FeatureOnlineStore.
Request message for [FeatureOnlineStoreAdminService.UpdateFeatureOnlineStore][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.UpdateFeatureOnlineStore].
Required. The FeatureOnlineStore's `name` field is used to identify the FeatureOnlineStore to be updated. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}`
Field mask is used to specify the fields to be overwritten in the FeatureOnlineStore resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `labels` * `description` * `bigtable` * `bigtable.auto_scaling` * `bigtable.enable_multi_region_replica`
Deletes a single FeatureOnlineStore. The FeatureOnlineStore must not contain any FeatureViews.
Request message for [FeatureOnlineStoreAdminService.DeleteFeatureOnlineStore][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.DeleteFeatureOnlineStore].
Required. The name of the FeatureOnlineStore to be deleted. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}`
If set to true, any FeatureViews and Features for this FeatureOnlineStore will also be deleted. (Otherwise, the request will only work if the FeatureOnlineStore has no FeatureViews.)
Creates a new FeatureView in a given FeatureOnlineStore.
Request message for [FeatureOnlineStoreAdminService.CreateFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.CreateFeatureView].
Required. The resource name of the FeatureOnlineStore to create FeatureViews. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}`
Required. The FeatureView to create.
Required. The ID to use for the FeatureView, which will become the final component of the FeatureView's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within a FeatureOnlineStore.
Immutable. If set to true, one on demand sync will be run immediately, regardless whether the [FeatureView.sync_config][google.cloud.aiplatform.v1.FeatureView.sync_config] is configured or not.
Gets details of a single FeatureView.
Request message for [FeatureOnlineStoreAdminService.GetFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.GetFeatureView].
Required. The name of the FeatureView resource. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Lists FeatureViews in a given FeatureOnlineStore.
Request message for [FeatureOnlineStoreAdminService.ListFeatureViews][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViews].
Required. The resource name of the FeatureOnlineStore to list FeatureViews. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}`
Lists the FeatureViews that match the filter expression. The following filters are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `>=`, and `<=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `>=`, and `<=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality as well as key presence. Examples: * `create_time > \"2020-01-31T15:30:00.000000Z\" OR update_time > \"2020-01-31T15:30:00.000000Z\"` --> FeatureViews created or updated after 2020-01-31T15:30:00.000000Z. * `labels.active = yes AND labels.env = prod` --> FeatureViews having both (active: yes) and (env: prod) labels. * `labels.env: *` --> Any FeatureView which has a label with 'env' as the key.
The maximum number of FeatureViews to return. The service may return fewer than this value. If unspecified, at most 1000 FeatureViews will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000.
A page token, received from a previous [FeatureOnlineStoreAdminService.ListFeatureViews][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViews] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeatureOnlineStoreAdminService.ListFeatureViews][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViews] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `feature_view_id` * `create_time` * `update_time`
Response message for [FeatureOnlineStoreAdminService.ListFeatureViews][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViews].
The FeatureViews matching the request.
A token, which can be sent as [ListFeatureViewsRequest.page_token][google.cloud.aiplatform.v1.ListFeatureViewsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Updates the parameters of a single FeatureView.
Request message for [FeatureOnlineStoreAdminService.UpdateFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.UpdateFeatureView].
Required. The FeatureView's `name` field is used to identify the FeatureView to be updated. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Field mask is used to specify the fields to be overwritten in the FeatureView resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `labels` * `service_agent_type` * `big_query_source` * `big_query_source.uri` * `big_query_source.entity_id_columns` * `feature_registry_source` * `feature_registry_source.feature_groups` * `sync_config` * `sync_config.cron` * `optimized_config.automatic_resources`
Deletes a single FeatureView.
Request message for [FeatureOnlineStoreAdminService.DeleteFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.DeleteFeatureView].
Required. The name of the FeatureView to be deleted. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Triggers on-demand sync for the FeatureView.
Request message for [FeatureOnlineStoreAdminService.SyncFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.SyncFeatureView].
Required. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Response message for [FeatureOnlineStoreAdminService.SyncFeatureView][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.SyncFeatureView].
Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}/featureViewSyncs/{feature_view_sync}`
Gets details of a single FeatureViewSync.
Request message for [FeatureOnlineStoreAdminService.GetFeatureViewSync][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.GetFeatureViewSync].
Required. The name of the FeatureViewSync resource. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}/featureViewSyncs/{feature_view_sync}`
Lists FeatureViewSyncs in a given FeatureView.
Request message for [FeatureOnlineStoreAdminService.ListFeatureViewSyncs][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViewSyncs].
Required. The resource name of the FeatureView to list FeatureViewSyncs. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Lists the FeatureViewSyncs that match the filter expression. The following filters are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `>=`, and `<=` comparisons. Values must be in RFC 3339 format. Examples: * `create_time > \"2020-01-31T15:30:00.000000Z\"` --> FeatureViewSyncs created after 2020-01-31T15:30:00.000000Z.
The maximum number of FeatureViewSyncs to return. The service may return fewer than this value. If unspecified, at most 1000 FeatureViewSyncs will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000.
A page token, received from a previous [FeatureOnlineStoreAdminService.ListFeatureViewSyncs][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViewSyncs] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeatureOnlineStoreAdminService.ListFeatureViewSyncs][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViewSyncs] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `create_time`
Response message for [FeatureOnlineStoreAdminService.ListFeatureViewSyncs][google.cloud.aiplatform.v1.FeatureOnlineStoreAdminService.ListFeatureViewSyncs].
The FeatureViewSyncs matching the request.
A token, which can be sent as [ListFeatureViewSyncsRequest.page_token][google.cloud.aiplatform.v1.ListFeatureViewSyncsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
A service for fetching feature values from the online store.
Fetch feature values under a FeatureView.
Request message for [FeatureOnlineStoreService.FetchFeatureValues][google.cloud.aiplatform.v1.FeatureOnlineStoreService.FetchFeatureValues]. All the features under the requested feature view will be returned.
Required. FeatureView resource format `projects/{project}/locations/{location}/featureOnlineStores/{featureOnlineStore}/featureViews/{featureView}`
Optional. The request key to fetch feature values for.
Optional. Response data format. If not set, [FeatureViewDataFormat.KEY_VALUE][google.cloud.aiplatform.v1.FeatureViewDataFormat.KEY_VALUE] will be used.
Search the nearest entities under a FeatureView. Search only works for indexable feature view; if a feature view isn't indexable, returns Invalid argument response.
The request message for [FeatureOnlineStoreService.SearchNearestEntities][google.cloud.aiplatform.v1.FeatureOnlineStoreService.SearchNearestEntities].
Required. FeatureView resource format `projects/{project}/locations/{location}/featureOnlineStores/{featureOnlineStore}/featureViews/{featureView}`
Required. The query.
Optional. If set to true, the full entities (including all vector values and metadata) of the nearest neighbors are returned; otherwise only entity id of the nearest neighbors will be returned. Note that returning full entities will significantly increase the latency and cost of the query.
Response message for [FeatureOnlineStoreService.SearchNearestEntities][google.cloud.aiplatform.v1.FeatureOnlineStoreService.SearchNearestEntities]
The nearest neighbors of the query entity.
The service that handles CRUD and List for resources for FeatureRegistry.
Creates a new FeatureGroup in a given project and location.
Request message for [FeatureRegistryService.CreateFeatureGroup][google.cloud.aiplatform.v1.FeatureRegistryService.CreateFeatureGroup].
Required. The resource name of the Location to create FeatureGroups. Format: `projects/{project}/locations/{location}`
Required. The FeatureGroup to create.
Required. The ID to use for this FeatureGroup, which will become the final component of the FeatureGroup's resource name. This value may be up to 128 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within the project and location.
Gets details of a single FeatureGroup.
Request message for [FeatureRegistryService.GetFeatureGroup][google.cloud.aiplatform.v1.FeatureRegistryService.GetFeatureGroup].
Required. The name of the FeatureGroup resource.
Lists FeatureGroups in a given project and location.
Request message for [FeatureRegistryService.ListFeatureGroups][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatureGroups].
Required. The resource name of the Location to list FeatureGroups. Format: `projects/{project}/locations/{location}`
Lists the FeatureGroups that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality and key presence. Examples: * `create_time > "2020-01-01" OR update_time > "2020-01-01"` FeatureGroups created or updated after 2020-01-01. * `labels.env = "prod"` FeatureGroups with label "env" set to "prod".
The maximum number of FeatureGroups to return. The service may return fewer than this value. If unspecified, at most 100 FeatureGroups will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
A page token, received from a previous [FeatureRegistryService.ListFeatureGroups][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatureGroups] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeatureRegistryService.ListFeatureGroups][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatureGroups] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields: * `create_time` * `update_time`
Response message for [FeatureRegistryService.ListFeatureGroups][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatureGroups].
The FeatureGroups matching the request.
A token, which can be sent as [ListFeatureGroupsRequest.page_token][google.cloud.aiplatform.v1.ListFeatureGroupsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Updates the parameters of a single FeatureGroup.
Request message for [FeatureRegistryService.UpdateFeatureGroup][google.cloud.aiplatform.v1.FeatureRegistryService.UpdateFeatureGroup].
Required. The FeatureGroup's `name` field is used to identify the FeatureGroup to be updated. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}`
Field mask is used to specify the fields to be overwritten in the FeatureGroup resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `labels` * `description` * `big_query` * `big_query.entity_id_columns`
Deletes a single FeatureGroup.
Request message for [FeatureRegistryService.DeleteFeatureGroup][google.cloud.aiplatform.v1.FeatureRegistryService.DeleteFeatureGroup].
Required. The name of the FeatureGroup to be deleted. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}`
If set to true, any Features under this FeatureGroup will also be deleted. (Otherwise, the request will only work if the FeatureGroup has no Features.)
Creates a new Feature in a given FeatureGroup.
Creates a batch of Features in a given FeatureGroup.
Gets details of a single Feature.
Lists Features in a given FeatureGroup.
Updates the parameters of a single Feature.
Deletes a single Feature.
A service for serving online feature values.
Reads Feature values of a specific entity of an EntityType. For reading feature values of multiple entities of an EntityType, please use StreamingReadFeatureValues.
Request message for [FeaturestoreOnlineServingService.ReadFeatureValues][google.cloud.aiplatform.v1.FeaturestoreOnlineServingService.ReadFeatureValues].
Required. The resource name of the EntityType for the entity being read. Value format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entityType}`. For example, for a machine learning model predicting user clicks on a website, an EntityType ID could be `user`.
Required. ID for a specific entity. For example, for a machine learning model predicting user clicks on a website, an entity ID could be `user_123`.
Required. Selector choosing Features of the target EntityType.
Reads Feature values for multiple entities. Depending on their size, data for different entities may be broken up across multiple responses.
Request message for [FeaturestoreOnlineServingService.StreamingReadFeatureValues][google.cloud.aiplatform.v1.FeaturestoreOnlineServingService.StreamingReadFeatureValues].
Required. The resource name of the entities' type. Value format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entityType}`. For example, for a machine learning model predicting user clicks on a website, an EntityType ID could be `user`.
Required. IDs of entities to read Feature values of. The maximum number of IDs is 100. For example, for a machine learning model predicting user clicks on a website, an entity ID could be `user_123`.
Required. Selector choosing Features of the target EntityType. Feature IDs will be deduplicated.
Writes Feature values of one or more entities of an EntityType. The Feature values are merged into existing entities if any. The Feature values to be written must have timestamp within the online storage retention.
Request message for [FeaturestoreOnlineServingService.WriteFeatureValues][google.cloud.aiplatform.v1.FeaturestoreOnlineServingService.WriteFeatureValues].
Required. The resource name of the EntityType for the entities being written. Value format: `projects/{project}/locations/{location}/featurestores/ {featurestore}/entityTypes/{entityType}`. For example, for a machine learning model predicting user clicks on a website, an EntityType ID could be `user`.
Required. The entities to be written. Up to 100,000 feature values can be written across all `payloads`.
Response message for [FeaturestoreOnlineServingService.WriteFeatureValues][google.cloud.aiplatform.v1.FeaturestoreOnlineServingService.WriteFeatureValues].
(message has no fields)
The service that handles CRUD and List for resources for Featurestore.
Creates a new Featurestore in a given project and location.
Request message for [FeaturestoreService.CreateFeaturestore][google.cloud.aiplatform.v1.FeaturestoreService.CreateFeaturestore].
Required. The resource name of the Location to create Featurestores. Format: `projects/{project}/locations/{location}`
Required. The Featurestore to create.
Required. The ID to use for this Featurestore, which will become the final component of the Featurestore's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within the project and location.
Gets details of a single Featurestore.
Request message for [FeaturestoreService.GetFeaturestore][google.cloud.aiplatform.v1.FeaturestoreService.GetFeaturestore].
Required. The name of the Featurestore resource.
Lists Featurestores in a given project and location.
Request message for [FeaturestoreService.ListFeaturestores][google.cloud.aiplatform.v1.FeaturestoreService.ListFeaturestores].
Required. The resource name of the Location to list Featurestores. Format: `projects/{project}/locations/{location}`
Lists the featurestores that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `online_serving_config.fixed_node_count`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. * `labels`: Supports key-value equality and key presence. Examples: * `create_time > "2020-01-01" OR update_time > "2020-01-01"` Featurestores created or updated after 2020-01-01. * `labels.env = "prod"` Featurestores with label "env" set to "prod".
The maximum number of Featurestores to return. The service may return fewer than this value. If unspecified, at most 100 Featurestores will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
A page token, received from a previous [FeaturestoreService.ListFeaturestores][google.cloud.aiplatform.v1.FeaturestoreService.ListFeaturestores] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeaturestoreService.ListFeaturestores][google.cloud.aiplatform.v1.FeaturestoreService.ListFeaturestores] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields: * `create_time` * `update_time` * `online_serving_config.fixed_node_count`
Mask specifying which fields to read.
Response message for [FeaturestoreService.ListFeaturestores][google.cloud.aiplatform.v1.FeaturestoreService.ListFeaturestores].
The Featurestores matching the request.
A token, which can be sent as [ListFeaturestoresRequest.page_token][google.cloud.aiplatform.v1.ListFeaturestoresRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Updates the parameters of a single Featurestore.
Request message for [FeaturestoreService.UpdateFeaturestore][google.cloud.aiplatform.v1.FeaturestoreService.UpdateFeaturestore].
Required. The Featurestore's `name` field is used to identify the Featurestore to be updated. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
Field mask is used to specify the fields to be overwritten in the Featurestore resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `labels` * `online_serving_config.fixed_node_count` * `online_serving_config.scaling` * `online_storage_ttl_days`
Deletes a single Featurestore. The Featurestore must not contain any EntityTypes or `force` must be set to true for the request to succeed.
Request message for [FeaturestoreService.DeleteFeaturestore][google.cloud.aiplatform.v1.FeaturestoreService.DeleteFeaturestore].
Required. The name of the Featurestore to be deleted. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
If set to true, any EntityTypes and Features for this Featurestore will also be deleted. (Otherwise, the request will only work if the Featurestore has no EntityTypes.)
Creates a new EntityType in a given Featurestore.
Request message for [FeaturestoreService.CreateEntityType][google.cloud.aiplatform.v1.FeaturestoreService.CreateEntityType].
Required. The resource name of the Featurestore to create EntityTypes. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
The EntityType to create.
Required. The ID to use for the EntityType, which will become the final component of the EntityType's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within a featurestore.
Gets details of a single EntityType.
Request message for [FeaturestoreService.GetEntityType][google.cloud.aiplatform.v1.FeaturestoreService.GetEntityType].
Required. The name of the EntityType resource. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}`
Lists EntityTypes in a given Featurestore.
Request message for [FeaturestoreService.ListEntityTypes][google.cloud.aiplatform.v1.FeaturestoreService.ListEntityTypes].
Required. The resource name of the Featurestore to list EntityTypes. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
Lists the EntityTypes that match the filter expression. The following filters are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `>=`, and `<=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `>=`, and `<=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality as well as key presence. Examples: * `create_time > \"2020-01-31T15:30:00.000000Z\" OR update_time > \"2020-01-31T15:30:00.000000Z\"` --> EntityTypes created or updated after 2020-01-31T15:30:00.000000Z. * `labels.active = yes AND labels.env = prod` --> EntityTypes having both (active: yes) and (env: prod) labels. * `labels.env: *` --> Any EntityType which has a label with 'env' as the key.
The maximum number of EntityTypes to return. The service may return fewer than this value. If unspecified, at most 1000 EntityTypes will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000.
A page token, received from a previous [FeaturestoreService.ListEntityTypes][google.cloud.aiplatform.v1.FeaturestoreService.ListEntityTypes] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeaturestoreService.ListEntityTypes][google.cloud.aiplatform.v1.FeaturestoreService.ListEntityTypes] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `entity_type_id` * `create_time` * `update_time`
Mask specifying which fields to read.
Response message for [FeaturestoreService.ListEntityTypes][google.cloud.aiplatform.v1.FeaturestoreService.ListEntityTypes].
The EntityTypes matching the request.
A token, which can be sent as [ListEntityTypesRequest.page_token][google.cloud.aiplatform.v1.ListEntityTypesRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Updates the parameters of a single EntityType.
Request message for [FeaturestoreService.UpdateEntityType][google.cloud.aiplatform.v1.FeaturestoreService.UpdateEntityType].
Required. The EntityType's `name` field is used to identify the EntityType to be updated. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}`
Field mask is used to specify the fields to be overwritten in the EntityType resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `description` * `labels` * `monitoring_config.snapshot_analysis.disabled` * `monitoring_config.snapshot_analysis.monitoring_interval_days` * `monitoring_config.snapshot_analysis.staleness_days` * `monitoring_config.import_features_analysis.state` * `monitoring_config.import_features_analysis.anomaly_detection_baseline` * `monitoring_config.numerical_threshold_config.value` * `monitoring_config.categorical_threshold_config.value` * `offline_storage_ttl_days`
Deletes a single EntityType. The EntityType must not have any Features or `force` must be set to true for the request to succeed.
Request message for [FeaturestoreService.DeleteEntityType][google.cloud.aiplatform.v1.FeaturestoreService.DeleteEntityType].
Required. The name of the EntityType to be deleted. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}`
If set to true, any Features for this EntityType will also be deleted. (Otherwise, the request will only work if the EntityType has no Features.)
Creates a new Feature in a given EntityType.
Creates a batch of Features in a given EntityType.
Gets details of a single Feature.
Lists Features in a given EntityType.
Updates the parameters of a single Feature.
Deletes a single Feature.
Imports Feature values into the Featurestore from a source storage. The progress of the import is tracked by the returned operation. The imported features are guaranteed to be visible to subsequent read operations after the operation is marked as successfully done. If an import operation fails, the Feature values returned from reads and exports may be inconsistent. If consistency is required, the caller must retry the same import request again and wait till the new operation returned is marked as successfully done. There are also scenarios where the caller can cause inconsistency. - Source data for import contains multiple distinct Feature values for the same entity ID and timestamp. - Source is modified during an import. This includes adding, updating, or removing source data and/or metadata. Examples of updating metadata include but are not limited to changing storage location, storage class, or retention policy. - Online serving cluster is under-provisioned.
Request message for [FeaturestoreService.ImportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ImportFeatureValues].
Details about the source data, including the location of the storage and the format.
Source of Feature timestamp for all Feature values of each entity. Timestamps must be millisecond-aligned.
Source column that holds the Feature timestamp for all Feature values in each entity.
Single Feature timestamp for all entities being imported. The timestamp must not have higher than millisecond precision.
Required. The resource name of the EntityType grouping the Features for which values are being imported. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entityType}`
Source column that holds entity IDs. If not provided, entity IDs are extracted from the column named entity_id.
Required. Specifications defining which Feature values to import from the entity. The request fails if no feature_specs are provided, and having multiple feature_specs for one Feature is not allowed.
If set, data will not be imported for online serving. This is typically used for backfilling, where Feature generation timestamps are not in the timestamp range needed for online serving.
Specifies the number of workers that are used to write data to the Featurestore. Consider the online serving capacity that you require to achieve the desired import throughput without interfering with online serving. The value must be positive, and less than or equal to 100. If not set, defaults to using 1 worker. The low count ensures minimal impact on online serving performance.
If true, API doesn't start ingestion analysis pipeline.
Batch reads Feature values from a Featurestore. This API enables batch reading Feature values, where each read instance in the batch may read Feature values of entities from one or more EntityTypes. Point-in-time correctness is guaranteed for Feature values of each read instance as of each instance's read timestamp.
Request message for [FeaturestoreService.BatchReadFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.BatchReadFeatureValues].
Each read instance consists of exactly one read timestamp and one or more entity IDs identifying entities of the corresponding EntityTypes whose Features are requested. Each output instance contains Feature values of requested entities concatenated together as of the read time. An example read instance may be `foo_entity_id, bar_entity_id, 2020-01-01T10:00:00.123Z`. An example output instance may be `foo_entity_id, bar_entity_id, 2020-01-01T10:00:00.123Z, foo_entity_feature1_value, bar_entity_feature2_value`. Timestamp in each read instance must be millisecond-aligned. `csv_read_instances` are read instances stored in a plain-text CSV file. The header should be: [ENTITY_TYPE_ID1], [ENTITY_TYPE_ID2], ..., timestamp The columns can be in any order. Values in the timestamp column must use the RFC 3339 format, e.g. `2012-07-30T10:43:17.123Z`.
Similar to csv_read_instances, but from BigQuery source.
Required. The resource name of the Featurestore from which to query Feature values. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
Required. Specifies output location and format.
When not empty, the specified fields in the *_read_instances source will be joined as-is in the output, in addition to those fields from the Featurestore Entity. For BigQuery source, the type of the pass-through values will be automatically inferred. For CSV source, the pass-through values will be passed as opaque bytes.
Required. Specifies EntityType grouping Features to read values of and settings.
Optional. Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision.
Exports Feature values from all the entities of a target EntityType.
Request message for [FeaturestoreService.ExportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ExportFeatureValues].
Required. The mode in which Feature values are exported.
Exports the latest Feature values of all entities of the EntityType within a time range.
Exports all historical values of all entities of the EntityType within a time range
Required. The resource name of the EntityType from which to export Feature values. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}`
Required. Specifies destination location and format.
Required. Selects Features to export values of.
Per-Feature export settings.
Delete Feature values from Featurestore. The progress of the deletion is tracked by the returned operation. The deleted feature values are guaranteed to be invisible to subsequent read operations after the operation is marked as successfully done. If a delete feature values operation fails, the feature values returned from reads and exports may be inconsistent. If consistency is required, the caller must retry the same delete request again and wait till the new operation returned is marked as successfully done.
Request message for [FeaturestoreService.DeleteFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.DeleteFeatureValues].
Defines options to select feature values to be deleted.
Select feature values to be deleted by specifying entities.
Select feature values to be deleted by specifying time range and features.
Required. The resource name of the EntityType grouping the Features for which values are being deleted from. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entityType}`
Searches Features matching a query in a given project.
Request message for [FeaturestoreService.SearchFeatures][google.cloud.aiplatform.v1.FeaturestoreService.SearchFeatures].
Required. The resource name of the Location to search Features. Format: `projects/{project}/locations/{location}`
Query string that is a conjunction of field-restricted queries and/or field-restricted filters. Field-restricted queries and filters can be combined using `AND` to form a conjunction. A field query is in the form FIELD:QUERY. This implicitly checks if QUERY exists as a substring within Feature's FIELD. The QUERY and the FIELD are converted to a sequence of words (i.e. tokens) for comparison. This is done by: * Removing leading/trailing whitespace and tokenizing the search value. Characters that are not one of alphanumeric `[a-zA-Z0-9]`, underscore `_`, or asterisk `*` are treated as delimiters for tokens. `*` is treated as a wildcard that matches characters within a token. * Ignoring case. * Prepending an asterisk to the first and appending an asterisk to the last token in QUERY. A QUERY must be either a singular token or a phrase. A phrase is one or multiple words enclosed in double quotation marks ("). With phrases, the order of the words is important. Words in the phrase must be matching in order and consecutively. Supported FIELDs for field-restricted queries: * `feature_id` * `description` * `entity_type_id` Examples: * `feature_id: foo` --> Matches a Feature with ID containing the substring `foo` (eg. `foo`, `foofeature`, `barfoo`). * `feature_id: foo*feature` --> Matches a Feature with ID containing the substring `foo*feature` (eg. `foobarfeature`). * `feature_id: foo AND description: bar` --> Matches a Feature with ID containing the substring `foo` and description containing the substring `bar`. Besides field queries, the following exact-match filters are supported. The exact-match filters do not support wildcards. Unlike field-restricted queries, exact-match filters are case-sensitive. * `feature_id`: Supports = comparisons. * `description`: Supports = comparisons. Multi-token filters should be enclosed in quotes. * `entity_type_id`: Supports = comparisons. * `value_type`: Supports = and != comparisons. * `labels`: Supports key-value equality as well as key presence. * `featurestore_id`: Supports = comparisons. Examples: * `description = "foo bar"` --> Any Feature with description exactly equal to `foo bar` * `value_type = DOUBLE` --> Features whose type is DOUBLE. * `labels.active = yes AND labels.env = prod` --> Features having both (active: yes) and (env: prod) labels. * `labels.env: *` --> Any Feature which has a label with `env` as the key.
The maximum number of Features to return. The service may return fewer than this value. If unspecified, at most 100 Features will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
A page token, received from a previous [FeaturestoreService.SearchFeatures][google.cloud.aiplatform.v1.FeaturestoreService.SearchFeatures] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeaturestoreService.SearchFeatures][google.cloud.aiplatform.v1.FeaturestoreService.SearchFeatures], except `page_size`, must match the call that provided the page token.
Response message for [FeaturestoreService.SearchFeatures][google.cloud.aiplatform.v1.FeaturestoreService.SearchFeatures].
The Features matching the request. Fields returned: * `name` * `description` * `labels` * `create_time` * `update_time`
A token, which can be sent as [SearchFeaturesRequest.page_token][google.cloud.aiplatform.v1.SearchFeaturesRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Service for managing Vertex AI's CachedContent resource.
Creates cached content, this call will initialize the cached content in the data storage, and users need to pay for the cache data storage.
Request message for [GenAiCacheService.CreateCachedContent][google.cloud.aiplatform.v1.GenAiCacheService.CreateCachedContent].
Required. The parent resource where the cached content will be created
Required. The cached content to create
Gets cached content configurations
Request message for [GenAiCacheService.GetCachedContent][google.cloud.aiplatform.v1.GenAiCacheService.GetCachedContent].
Required. The resource name referring to the cached content
Updates cached content configurations
Request message for [GenAiCacheService.UpdateCachedContent][google.cloud.aiplatform.v1.GenAiCacheService.UpdateCachedContent]. Only expire_time or ttl can be updated.
Required. The cached content to update
Required. The list of fields to update.
Deletes cached content
Request message for [GenAiCacheService.DeleteCachedContent][google.cloud.aiplatform.v1.GenAiCacheService.DeleteCachedContent].
Required. The resource name referring to the cached content
Lists cached contents in a project
Request to list CachedContents.
Required. The parent, which owns this collection of cached contents.
Optional. The maximum number of cached contents to return. The service may return fewer than this value. If unspecified, some default (under maximum) number of items will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000.
Optional. A page token, received from a previous `ListCachedContents` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListCachedContents` must match the call that provided the page token.
Response with a list of CachedContents.
List of cached contents.
A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages.
A service for creating and managing GenAI Tuning Jobs.
Creates a TuningJob. A created TuningJob right away will be attempted to be run.
Request message for [GenAiTuningService.CreateTuningJob][google.cloud.aiplatform.v1.GenAiTuningService.CreateTuningJob].
Required. The resource name of the Location to create the TuningJob in. Format: `projects/{project}/locations/{location}`
Required. The TuningJob to create.
Gets a TuningJob.
Request message for [GenAiTuningService.GetTuningJob][google.cloud.aiplatform.v1.GenAiTuningService.GetTuningJob].
Required. The name of the TuningJob resource. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}`
Lists TuningJobs in a Location.
Request message for [GenAiTuningService.ListTuningJobs][google.cloud.aiplatform.v1.GenAiTuningService.ListTuningJobs].
Required. The resource name of the Location to list the TuningJobs from. Format: `projects/{project}/locations/{location}`
Optional. The standard list filter.
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListTuningJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListTuningJobsResponse.next_page_token] of the previous GenAiTuningService.ListTuningJob][] call.
Response message for [GenAiTuningService.ListTuningJobs][google.cloud.aiplatform.v1.GenAiTuningService.ListTuningJobs]
List of TuningJobs in the requested page.
A token to retrieve the next page of results. Pass to [ListTuningJobsRequest.page_token][google.cloud.aiplatform.v1.ListTuningJobsRequest.page_token] to obtain that page.
Cancels a TuningJob. Starts asynchronous cancellation on the TuningJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use [GenAiTuningService.GetTuningJob][google.cloud.aiplatform.v1.GenAiTuningService.GetTuningJob] or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On successful cancellation, the TuningJob is not deleted; instead it becomes a job with a [TuningJob.error][google.cloud.aiplatform.v1.TuningJob.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [TuningJob.state][google.cloud.aiplatform.v1.TuningJob.state] is set to `CANCELLED`.
Request message for [GenAiTuningService.CancelTuningJob][google.cloud.aiplatform.v1.GenAiTuningService.CancelTuningJob].
Required. The name of the TuningJob to cancel. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}`
Rebase a TunedModel.
Request message for [GenAiTuningService.RebaseTunedModel][google.cloud.aiplatform.v1.GenAiTuningService.RebaseTunedModel].
Required. The resource name of the Location into which to rebase the Model. Format: `projects/{project}/locations/{location}`
Required. TunedModel reference to retrieve the legacy model information.
Optional. The TuningJob to be updated. Users can use this TuningJob field to overwrite tuning configs.
Optional. The Google Cloud Storage location to write the artifacts.
Optional. By default, bison to gemini migration will always create new model/endpoint, but for gemini-1.0 to gemini-1.5 migration, we default deploy to the same endpoint. See details in this Section.
A service for managing Vertex AI's IndexEndpoints.
Creates an IndexEndpoint.
Request message for [IndexEndpointService.CreateIndexEndpoint][google.cloud.aiplatform.v1.IndexEndpointService.CreateIndexEndpoint].
Required. The resource name of the Location to create the IndexEndpoint in. Format: `projects/{project}/locations/{location}`
Required. The IndexEndpoint to create.
Gets an IndexEndpoint.
Request message for [IndexEndpointService.GetIndexEndpoint][google.cloud.aiplatform.v1.IndexEndpointService.GetIndexEndpoint]
Required. The name of the IndexEndpoint resource. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
Lists IndexEndpoints in a Location.
Request message for [IndexEndpointService.ListIndexEndpoints][google.cloud.aiplatform.v1.IndexEndpointService.ListIndexEndpoints].
Required. The resource name of the Location from which to list the IndexEndpoints. Format: `projects/{project}/locations/{location}`
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `index_endpoint` supports = and !=. `index_endpoint` represents the IndexEndpoint ID, ie. the last segment of the IndexEndpoint's [resourcename][google.cloud.aiplatform.v1.IndexEndpoint.name]. * `display_name` supports =, != and regex() (uses [re2](https://github.com/google/re2/wiki/Syntax) syntax) * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* or labels:key - key existence A key including a space must be quoted. `labels."a key"`. Some examples: * `index_endpoint="1"` * `display_name="myDisplayName"` * `regex(display_name, "^A") -> The display name starts with an A. * `labels.myKey="myValue"`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListIndexEndpointsResponse.next_page_token][google.cloud.aiplatform.v1.ListIndexEndpointsResponse.next_page_token] of the previous [IndexEndpointService.ListIndexEndpoints][google.cloud.aiplatform.v1.IndexEndpointService.ListIndexEndpoints] call.
Optional. Mask specifying which fields to read.
Response message for [IndexEndpointService.ListIndexEndpoints][google.cloud.aiplatform.v1.IndexEndpointService.ListIndexEndpoints].
List of IndexEndpoints in the requested page.
A token to retrieve next page of results. Pass to [ListIndexEndpointsRequest.page_token][google.cloud.aiplatform.v1.ListIndexEndpointsRequest.page_token] to obtain that page.
Updates an IndexEndpoint.
Request message for [IndexEndpointService.UpdateIndexEndpoint][google.cloud.aiplatform.v1.IndexEndpointService.UpdateIndexEndpoint].
Required. The IndexEndpoint which replaces the resource on the server.
Required. The update mask applies to the resource. See [google.protobuf.FieldMask][google.protobuf.FieldMask].
Deletes an IndexEndpoint.
Request message for [IndexEndpointService.DeleteIndexEndpoint][google.cloud.aiplatform.v1.IndexEndpointService.DeleteIndexEndpoint].
Required. The name of the IndexEndpoint resource to be deleted. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
Deploys an Index into this IndexEndpoint, creating a DeployedIndex within it. Only non-empty Indexes can be deployed.
Request message for [IndexEndpointService.DeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.DeployIndex].
Required. The name of the IndexEndpoint resource into which to deploy an Index. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
Required. The DeployedIndex to be created within the IndexEndpoint.
Undeploys an Index from an IndexEndpoint, removing a DeployedIndex from it, and freeing all resources it's using.
Request message for [IndexEndpointService.UndeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.UndeployIndex].
Required. The name of the IndexEndpoint resource from which to undeploy an Index. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
Required. The ID of the DeployedIndex to be undeployed from the IndexEndpoint.
Update an existing DeployedIndex under an IndexEndpoint.
Request message for [IndexEndpointService.MutateDeployedIndex][google.cloud.aiplatform.v1.IndexEndpointService.MutateDeployedIndex].
Required. The name of the IndexEndpoint resource into which to deploy an Index. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
Required. The DeployedIndex to be updated within the IndexEndpoint. Currently, the updatable fields are [DeployedIndex.automatic_resources][google.cloud.aiplatform.v1.DeployedIndex.automatic_resources] and [DeployedIndex.dedicated_resources][google.cloud.aiplatform.v1.DeployedIndex.dedicated_resources]
A service for creating and managing Vertex AI's Index resources.
Creates an Index.
Request message for [IndexService.CreateIndex][google.cloud.aiplatform.v1.IndexService.CreateIndex].
Required. The resource name of the Location to create the Index in. Format: `projects/{project}/locations/{location}`
Required. The Index to create.
Gets an Index.
Request message for [IndexService.GetIndex][google.cloud.aiplatform.v1.IndexService.GetIndex]
Required. The name of the Index resource. Format: `projects/{project}/locations/{location}/indexes/{index}`
Lists Indexes in a Location.
Request message for [IndexService.ListIndexes][google.cloud.aiplatform.v1.IndexService.ListIndexes].
Required. The resource name of the Location from which to list the Indexes. Format: `projects/{project}/locations/{location}`
The standard list filter.
The standard list page size.
The standard list page token. Typically obtained via [ListIndexesResponse.next_page_token][google.cloud.aiplatform.v1.ListIndexesResponse.next_page_token] of the previous [IndexService.ListIndexes][google.cloud.aiplatform.v1.IndexService.ListIndexes] call.
Mask specifying which fields to read.
Response message for [IndexService.ListIndexes][google.cloud.aiplatform.v1.IndexService.ListIndexes].
List of indexes in the requested page.
A token to retrieve next page of results. Pass to [ListIndexesRequest.page_token][google.cloud.aiplatform.v1.ListIndexesRequest.page_token] to obtain that page.
Updates an Index.
Request message for [IndexService.UpdateIndex][google.cloud.aiplatform.v1.IndexService.UpdateIndex].
Required. The Index which updates the resource on the server.
The update mask applies to the resource. For the `FieldMask` definition, see [google.protobuf.FieldMask][google.protobuf.FieldMask].
Deletes an Index. An Index can only be deleted when all its [DeployedIndexes][google.cloud.aiplatform.v1.Index.deployed_indexes] had been undeployed.
Request message for [IndexService.DeleteIndex][google.cloud.aiplatform.v1.IndexService.DeleteIndex].
Required. The name of the Index resource to be deleted. Format: `projects/{project}/locations/{location}/indexes/{index}`
Add/update Datapoints into an Index.
Request message for [IndexService.UpsertDatapoints][google.cloud.aiplatform.v1.IndexService.UpsertDatapoints]
Required. The name of the Index resource to be updated. Format: `projects/{project}/locations/{location}/indexes/{index}`
A list of datapoints to be created/updated.
Optional. Update mask is used to specify the fields to be overwritten in the datapoints by the update. The fields specified in the update_mask are relative to each IndexDatapoint inside datapoints, not the full request. Updatable fields: * Use `all_restricts` to update both restricts and numeric_restricts.
Response message for [IndexService.UpsertDatapoints][google.cloud.aiplatform.v1.IndexService.UpsertDatapoints]
(message has no fields)
Remove Datapoints from an Index.
Request message for [IndexService.RemoveDatapoints][google.cloud.aiplatform.v1.IndexService.RemoveDatapoints]
Required. The name of the Index resource to be updated. Format: `projects/{project}/locations/{location}/indexes/{index}`
A list of datapoint ids to be deleted.
Response message for [IndexService.RemoveDatapoints][google.cloud.aiplatform.v1.IndexService.RemoveDatapoints]
(message has no fields)
A service for creating and managing Vertex AI's jobs.
Creates a CustomJob. A created CustomJob right away will be attempted to be run.
Request message for [JobService.CreateCustomJob][google.cloud.aiplatform.v1.JobService.CreateCustomJob].
Required. The resource name of the Location to create the CustomJob in. Format: `projects/{project}/locations/{location}`
Required. The CustomJob to create.
Gets a CustomJob.
Request message for [JobService.GetCustomJob][google.cloud.aiplatform.v1.JobService.GetCustomJob].
Required. The name of the CustomJob resource. Format: `projects/{project}/locations/{location}/customJobs/{custom_job}`
Lists CustomJobs in a Location.
Request message for [JobService.ListCustomJobs][google.cloud.aiplatform.v1.JobService.ListCustomJobs].
Required. The resource name of the Location to list the CustomJobs from. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token. Typically obtained via [ListCustomJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListCustomJobsResponse.next_page_token] of the previous [JobService.ListCustomJobs][google.cloud.aiplatform.v1.JobService.ListCustomJobs] call.
Mask specifying which fields to read.
Response message for [JobService.ListCustomJobs][google.cloud.aiplatform.v1.JobService.ListCustomJobs]
List of CustomJobs in the requested page.
A token to retrieve the next page of results. Pass to [ListCustomJobsRequest.page_token][google.cloud.aiplatform.v1.ListCustomJobsRequest.page_token] to obtain that page.
Deletes a CustomJob.
Request message for [JobService.DeleteCustomJob][google.cloud.aiplatform.v1.JobService.DeleteCustomJob].
Required. The name of the CustomJob resource to be deleted. Format: `projects/{project}/locations/{location}/customJobs/{custom_job}`
Cancels a CustomJob. Starts asynchronous cancellation on the CustomJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use [JobService.GetCustomJob][google.cloud.aiplatform.v1.JobService.GetCustomJob] or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On successful cancellation, the CustomJob is not deleted; instead it becomes a job with a [CustomJob.error][google.cloud.aiplatform.v1.CustomJob.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [CustomJob.state][google.cloud.aiplatform.v1.CustomJob.state] is set to `CANCELLED`.
Request message for [JobService.CancelCustomJob][google.cloud.aiplatform.v1.JobService.CancelCustomJob].
Required. The name of the CustomJob to cancel. Format: `projects/{project}/locations/{location}/customJobs/{custom_job}`
Creates a DataLabelingJob.
Request message for [JobService.CreateDataLabelingJob][google.cloud.aiplatform.v1.JobService.CreateDataLabelingJob].
Required. The parent of the DataLabelingJob. Format: `projects/{project}/locations/{location}`
Required. The DataLabelingJob to create.
Gets a DataLabelingJob.
Request message for [JobService.GetDataLabelingJob][google.cloud.aiplatform.v1.JobService.GetDataLabelingJob].
Required. The name of the DataLabelingJob. Format: `projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`
Lists DataLabelingJobs in a Location.
Request message for [JobService.ListDataLabelingJobs][google.cloud.aiplatform.v1.JobService.ListDataLabelingJobs].
Required. The parent of the DataLabelingJob. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token.
Mask specifying which fields to read. FieldMask represents a set of symbolic field paths. For example, the mask can be `paths: "name"`. The "name" here is a field in DataLabelingJob. If this field is not set, all fields of the DataLabelingJob are returned.
A comma-separated list of fields to order by, sorted in ascending order by default. Use `desc` after a field name for descending.
Response message for [JobService.ListDataLabelingJobs][google.cloud.aiplatform.v1.JobService.ListDataLabelingJobs].
A list of DataLabelingJobs that matches the specified filter in the request.
The standard List next-page token.
Deletes a DataLabelingJob.
Request message for [JobService.DeleteDataLabelingJob][google.cloud.aiplatform.v1.JobService.DeleteDataLabelingJob].
Required. The name of the DataLabelingJob to be deleted. Format: `projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`
Cancels a DataLabelingJob. Success of cancellation is not guaranteed.
Request message for [JobService.CancelDataLabelingJob][google.cloud.aiplatform.v1.JobService.CancelDataLabelingJob].
Required. The name of the DataLabelingJob. Format: `projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`
Creates a HyperparameterTuningJob
Request message for [JobService.CreateHyperparameterTuningJob][google.cloud.aiplatform.v1.JobService.CreateHyperparameterTuningJob].
Required. The resource name of the Location to create the HyperparameterTuningJob in. Format: `projects/{project}/locations/{location}`
Required. The HyperparameterTuningJob to create.
Gets a HyperparameterTuningJob
Request message for [JobService.GetHyperparameterTuningJob][google.cloud.aiplatform.v1.JobService.GetHyperparameterTuningJob].
Required. The name of the HyperparameterTuningJob resource. Format: `projects/{project}/locations/{location}/hyperparameterTuningJobs/{hyperparameter_tuning_job}`
Lists HyperparameterTuningJobs in a Location.
Request message for [JobService.ListHyperparameterTuningJobs][google.cloud.aiplatform.v1.JobService.ListHyperparameterTuningJobs].
Required. The resource name of the Location to list the HyperparameterTuningJobs from. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token. Typically obtained via [ListHyperparameterTuningJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListHyperparameterTuningJobsResponse.next_page_token] of the previous [JobService.ListHyperparameterTuningJobs][google.cloud.aiplatform.v1.JobService.ListHyperparameterTuningJobs] call.
Mask specifying which fields to read.
Response message for [JobService.ListHyperparameterTuningJobs][google.cloud.aiplatform.v1.JobService.ListHyperparameterTuningJobs]
List of HyperparameterTuningJobs in the requested page. [HyperparameterTuningJob.trials][google.cloud.aiplatform.v1.HyperparameterTuningJob.trials] of the jobs will be not be returned.
A token to retrieve the next page of results. Pass to [ListHyperparameterTuningJobsRequest.page_token][google.cloud.aiplatform.v1.ListHyperparameterTuningJobsRequest.page_token] to obtain that page.
Deletes a HyperparameterTuningJob.
Request message for [JobService.DeleteHyperparameterTuningJob][google.cloud.aiplatform.v1.JobService.DeleteHyperparameterTuningJob].
Required. The name of the HyperparameterTuningJob resource to be deleted. Format: `projects/{project}/locations/{location}/hyperparameterTuningJobs/{hyperparameter_tuning_job}`
Cancels a HyperparameterTuningJob. Starts asynchronous cancellation on the HyperparameterTuningJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use [JobService.GetHyperparameterTuningJob][google.cloud.aiplatform.v1.JobService.GetHyperparameterTuningJob] or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On successful cancellation, the HyperparameterTuningJob is not deleted; instead it becomes a job with a [HyperparameterTuningJob.error][google.cloud.aiplatform.v1.HyperparameterTuningJob.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [HyperparameterTuningJob.state][google.cloud.aiplatform.v1.HyperparameterTuningJob.state] is set to `CANCELLED`.
Request message for [JobService.CancelHyperparameterTuningJob][google.cloud.aiplatform.v1.JobService.CancelHyperparameterTuningJob].
Required. The name of the HyperparameterTuningJob to cancel. Format: `projects/{project}/locations/{location}/hyperparameterTuningJobs/{hyperparameter_tuning_job}`
Creates a NasJob
Request message for [JobService.CreateNasJob][google.cloud.aiplatform.v1.JobService.CreateNasJob].
Required. The resource name of the Location to create the NasJob in. Format: `projects/{project}/locations/{location}`
Required. The NasJob to create.
Gets a NasJob
Request message for [JobService.GetNasJob][google.cloud.aiplatform.v1.JobService.GetNasJob].
Required. The name of the NasJob resource. Format: `projects/{project}/locations/{location}/nasJobs/{nas_job}`
Lists NasJobs in a Location.
Request message for [JobService.ListNasJobs][google.cloud.aiplatform.v1.JobService.ListNasJobs].
Required. The resource name of the Location to list the NasJobs from. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token. Typically obtained via [ListNasJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListNasJobsResponse.next_page_token] of the previous [JobService.ListNasJobs][google.cloud.aiplatform.v1.JobService.ListNasJobs] call.
Mask specifying which fields to read.
Response message for [JobService.ListNasJobs][google.cloud.aiplatform.v1.JobService.ListNasJobs]
List of NasJobs in the requested page. [NasJob.nas_job_output][google.cloud.aiplatform.v1.NasJob.nas_job_output] of the jobs will not be returned.
A token to retrieve the next page of results. Pass to [ListNasJobsRequest.page_token][google.cloud.aiplatform.v1.ListNasJobsRequest.page_token] to obtain that page.
Deletes a NasJob.
Request message for [JobService.DeleteNasJob][google.cloud.aiplatform.v1.JobService.DeleteNasJob].
Required. The name of the NasJob resource to be deleted. Format: `projects/{project}/locations/{location}/nasJobs/{nas_job}`
Cancels a NasJob. Starts asynchronous cancellation on the NasJob. The server makes a best effort to cancel the job, but success is not guaranteed. Clients can use [JobService.GetNasJob][google.cloud.aiplatform.v1.JobService.GetNasJob] or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On successful cancellation, the NasJob is not deleted; instead it becomes a job with a [NasJob.error][google.cloud.aiplatform.v1.NasJob.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [NasJob.state][google.cloud.aiplatform.v1.NasJob.state] is set to `CANCELLED`.
Request message for [JobService.CancelNasJob][google.cloud.aiplatform.v1.JobService.CancelNasJob].
Required. The name of the NasJob to cancel. Format: `projects/{project}/locations/{location}/nasJobs/{nas_job}`
Gets a NasTrialDetail.
Request message for [JobService.GetNasTrialDetail][google.cloud.aiplatform.v1.JobService.GetNasTrialDetail].
Required. The name of the NasTrialDetail resource. Format: `projects/{project}/locations/{location}/nasJobs/{nas_job}/nasTrialDetails/{nas_trial_detail}`
List top NasTrialDetails of a NasJob.
Request message for [JobService.ListNasTrialDetails][google.cloud.aiplatform.v1.JobService.ListNasTrialDetails].
Required. The name of the NasJob resource. Format: `projects/{project}/locations/{location}/nasJobs/{nas_job}`
The standard list page size.
The standard list page token. Typically obtained via [ListNasTrialDetailsResponse.next_page_token][google.cloud.aiplatform.v1.ListNasTrialDetailsResponse.next_page_token] of the previous [JobService.ListNasTrialDetails][google.cloud.aiplatform.v1.JobService.ListNasTrialDetails] call.
Response message for [JobService.ListNasTrialDetails][google.cloud.aiplatform.v1.JobService.ListNasTrialDetails]
List of top NasTrials in the requested page.
A token to retrieve the next page of results. Pass to [ListNasTrialDetailsRequest.page_token][google.cloud.aiplatform.v1.ListNasTrialDetailsRequest.page_token] to obtain that page.
Creates a BatchPredictionJob. A BatchPredictionJob once created will right away be attempted to start.
Request message for [JobService.CreateBatchPredictionJob][google.cloud.aiplatform.v1.JobService.CreateBatchPredictionJob].
Required. The resource name of the Location to create the BatchPredictionJob in. Format: `projects/{project}/locations/{location}`
Required. The BatchPredictionJob to create.
Gets a BatchPredictionJob
Request message for [JobService.GetBatchPredictionJob][google.cloud.aiplatform.v1.JobService.GetBatchPredictionJob].
Required. The name of the BatchPredictionJob resource. Format: `projects/{project}/locations/{location}/batchPredictionJobs/{batch_prediction_job}`
Lists BatchPredictionJobs in a Location.
Request message for [JobService.ListBatchPredictionJobs][google.cloud.aiplatform.v1.JobService.ListBatchPredictionJobs].
Required. The resource name of the Location to list the BatchPredictionJobs from. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `model_display_name` supports `=`, `!=` comparisons. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token. Typically obtained via [ListBatchPredictionJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListBatchPredictionJobsResponse.next_page_token] of the previous [JobService.ListBatchPredictionJobs][google.cloud.aiplatform.v1.JobService.ListBatchPredictionJobs] call.
Mask specifying which fields to read.
Response message for [JobService.ListBatchPredictionJobs][google.cloud.aiplatform.v1.JobService.ListBatchPredictionJobs]
List of BatchPredictionJobs in the requested page.
A token to retrieve the next page of results. Pass to [ListBatchPredictionJobsRequest.page_token][google.cloud.aiplatform.v1.ListBatchPredictionJobsRequest.page_token] to obtain that page.
Deletes a BatchPredictionJob. Can only be called on jobs that already finished.
Request message for [JobService.DeleteBatchPredictionJob][google.cloud.aiplatform.v1.JobService.DeleteBatchPredictionJob].
Required. The name of the BatchPredictionJob resource to be deleted. Format: `projects/{project}/locations/{location}/batchPredictionJobs/{batch_prediction_job}`
Cancels a BatchPredictionJob. Starts asynchronous cancellation on the BatchPredictionJob. The server makes the best effort to cancel the job, but success is not guaranteed. Clients can use [JobService.GetBatchPredictionJob][google.cloud.aiplatform.v1.JobService.GetBatchPredictionJob] or other methods to check whether the cancellation succeeded or whether the job completed despite cancellation. On a successful cancellation, the BatchPredictionJob is not deleted;instead its [BatchPredictionJob.state][google.cloud.aiplatform.v1.BatchPredictionJob.state] is set to `CANCELLED`. Any files already outputted by the job are not deleted.
Request message for [JobService.CancelBatchPredictionJob][google.cloud.aiplatform.v1.JobService.CancelBatchPredictionJob].
Required. The name of the BatchPredictionJob to cancel. Format: `projects/{project}/locations/{location}/batchPredictionJobs/{batch_prediction_job}`
Creates a ModelDeploymentMonitoringJob. It will run periodically on a configured interval.
Request message for [JobService.CreateModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.CreateModelDeploymentMonitoringJob].
Required. The parent of the ModelDeploymentMonitoringJob. Format: `projects/{project}/locations/{location}`
Required. The ModelDeploymentMonitoringJob to create
Searches Model Monitoring Statistics generated within a given time window.
Request message for [JobService.SearchModelDeploymentMonitoringStatsAnomalies][google.cloud.aiplatform.v1.JobService.SearchModelDeploymentMonitoringStatsAnomalies].
Required. ModelDeploymentMonitoring Job resource name. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Required. The DeployedModel ID of the [ModelDeploymentMonitoringObjectiveConfig.deployed_model_id].
The feature display name. If specified, only return the stats belonging to this feature. Format: [ModelMonitoringStatsAnomalies.FeatureHistoricStatsAnomalies.feature_display_name][google.cloud.aiplatform.v1.ModelMonitoringStatsAnomalies.FeatureHistoricStatsAnomalies.feature_display_name], example: "user_destination".
Required. Objectives of the stats to retrieve.
The standard list page size.
A page token received from a previous [JobService.SearchModelDeploymentMonitoringStatsAnomalies][google.cloud.aiplatform.v1.JobService.SearchModelDeploymentMonitoringStatsAnomalies] call.
The earliest timestamp of stats being generated. If not set, indicates fetching stats till the earliest possible one.
The latest timestamp of stats being generated. If not set, indicates feching stats till the latest possible one.
Response message for [JobService.SearchModelDeploymentMonitoringStatsAnomalies][google.cloud.aiplatform.v1.JobService.SearchModelDeploymentMonitoringStatsAnomalies].
Stats retrieved for requested objectives. There are at most 1000 [ModelMonitoringStatsAnomalies.FeatureHistoricStatsAnomalies.prediction_stats][google.cloud.aiplatform.v1.ModelMonitoringStatsAnomalies.FeatureHistoricStatsAnomalies.prediction_stats] in the response.
The page token that can be used by the next [JobService.SearchModelDeploymentMonitoringStatsAnomalies][google.cloud.aiplatform.v1.JobService.SearchModelDeploymentMonitoringStatsAnomalies] call.
Gets a ModelDeploymentMonitoringJob.
Request message for [JobService.GetModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.GetModelDeploymentMonitoringJob].
Required. The resource name of the ModelDeploymentMonitoringJob. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Lists ModelDeploymentMonitoringJobs in a Location.
Request message for [JobService.ListModelDeploymentMonitoringJobs][google.cloud.aiplatform.v1.JobService.ListModelDeploymentMonitoringJobs].
Required. The parent of the ModelDeploymentMonitoringJob. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="JOB_STATE_SUCCEEDED" AND display_name:"my_job_*"` * `state!="JOB_STATE_FAILED" OR display_name="my_job"` * `NOT display_name="my_job"` * `create_time>"2021-05-18T00:00:00Z"` * `labels.keyA=valueA` * `labels.keyB:*`
The standard list page size.
The standard list page token.
Mask specifying which fields to read
Response message for [JobService.ListModelDeploymentMonitoringJobs][google.cloud.aiplatform.v1.JobService.ListModelDeploymentMonitoringJobs].
A list of ModelDeploymentMonitoringJobs that matches the specified filter in the request.
The standard List next-page token.
Updates a ModelDeploymentMonitoringJob.
Request message for [JobService.UpdateModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.UpdateModelDeploymentMonitoringJob].
Required. The model monitoring configuration which replaces the resource on the server.
Required. The update mask is used to specify the fields to be overwritten in the ModelDeploymentMonitoringJob resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. For the objective config, the user can either provide the update mask for model_deployment_monitoring_objective_configs or any combination of its nested fields, such as: model_deployment_monitoring_objective_configs.objective_config.training_dataset. Updatable fields: * `display_name` * `model_deployment_monitoring_schedule_config` * `model_monitoring_alert_config` * `logging_sampling_strategy` * `labels` * `log_ttl` * `enable_monitoring_pipeline_logs` . and * `model_deployment_monitoring_objective_configs` . or * `model_deployment_monitoring_objective_configs.objective_config.training_dataset` * `model_deployment_monitoring_objective_configs.objective_config.training_prediction_skew_detection_config` * `model_deployment_monitoring_objective_configs.objective_config.prediction_drift_detection_config`
Deletes a ModelDeploymentMonitoringJob.
Request message for [JobService.DeleteModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.DeleteModelDeploymentMonitoringJob].
Required. The resource name of the model monitoring job to delete. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Pauses a ModelDeploymentMonitoringJob. If the job is running, the server makes a best effort to cancel the job. Will mark [ModelDeploymentMonitoringJob.state][google.cloud.aiplatform.v1.ModelDeploymentMonitoringJob.state] to 'PAUSED'.
Request message for [JobService.PauseModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.PauseModelDeploymentMonitoringJob].
Required. The resource name of the ModelDeploymentMonitoringJob to pause. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Resumes a paused ModelDeploymentMonitoringJob. It will start to run from next scheduled time. A deleted ModelDeploymentMonitoringJob can't be resumed.
Request message for [JobService.ResumeModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.ResumeModelDeploymentMonitoringJob].
Required. The resource name of the ModelDeploymentMonitoringJob to resume. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Service for LLM related utility functions.
Perform a token counting.
Request message for [PredictionService.CountTokens][].
Required. The name of the Endpoint requested to perform token counting. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Optional. The name of the publisher model requested to serve the prediction. Format: `projects/{project}/locations/{location}/publishers/*/models/*`
Optional. The instances that are the input to token counting call. Schema is identical to the prediction schema of the underlying model.
Optional. Input content.
Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph.
Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model.
Optional. Generation config that the model will use to generate the response.
Response message for [PredictionService.CountTokens][].
The total number of tokens counted across all instances from the request.
The total number of billable characters counted across all instances from the request.
Output only. List of modalities that were processed in the request input.
Return a list of tokens based on the input text.
Request message for ComputeTokens RPC call.
Required. The name of the Endpoint requested to get lists of tokens and token ids.
Optional. The instances that are the input to token computing API call. Schema is identical to the prediction schema of the text model, even for the non-text models, like chat models, or Codey models.
Optional. The name of the publisher model requested to serve the prediction. Format: projects/{project}/locations/{location}/publishers/*/models/*
Optional. Input content.
Response message for ComputeTokens RPC call.
Lists of tokens info from the input. A ComputeTokensRequest could have multiple instances with a prompt in each instance. We also need to return lists of tokens info for the request with multiple instances.
MatchService is a Google managed service for efficient vector similarity search at scale.
Finds the nearest neighbors of each vector within the request.
The request message for [MatchService.FindNeighbors][google.cloud.aiplatform.v1.MatchService.FindNeighbors].
Required. The name of the index endpoint. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
The ID of the DeployedIndex that will serve the request. This request is sent to a specific IndexEndpoint, as per the IndexEndpoint.network. That IndexEndpoint also has IndexEndpoint.deployed_indexes, and each such index has a DeployedIndex.id field. The value of the field below must equal one of the DeployedIndex.id fields of the IndexEndpoint that is being called for this request.
The list of queries.
If set to true, the full datapoints (including all vector values and restricts) of the nearest neighbors are returned. Note that returning full datapoint will significantly increase the latency and cost of the query.
The response message for [MatchService.FindNeighbors][google.cloud.aiplatform.v1.MatchService.FindNeighbors].
The nearest neighbors of the query datapoints.
Reads the datapoints/vectors of the given IDs. A maximum of 1000 datapoints can be retrieved in a batch.
The request message for [MatchService.ReadIndexDatapoints][google.cloud.aiplatform.v1.MatchService.ReadIndexDatapoints].
Required. The name of the index endpoint. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
The ID of the DeployedIndex that will serve the request.
IDs of the datapoints to be searched for.
The response message for [MatchService.ReadIndexDatapoints][google.cloud.aiplatform.v1.MatchService.ReadIndexDatapoints].
The result list of datapoints.
Service for reading and writing metadata entries.
Initializes a MetadataStore, including allocation of resources.
Request message for [MetadataService.CreateMetadataStore][google.cloud.aiplatform.v1.MetadataService.CreateMetadataStore].
Required. The resource name of the Location where the MetadataStore should be created. Format: `projects/{project}/locations/{location}/`
Required. The MetadataStore to create.
The {metadatastore} portion of the resource name with the format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}` If not provided, the MetadataStore's ID will be a UUID generated by the service. Must be 4-128 characters in length. Valid characters are `/[a-z][0-9]-/`. Must be unique across all MetadataStores in the parent Location. (Otherwise the request will fail with ALREADY_EXISTS, or PERMISSION_DENIED if the caller can't view the preexisting MetadataStore.)
Retrieves a specific MetadataStore.
Request message for [MetadataService.GetMetadataStore][google.cloud.aiplatform.v1.MetadataService.GetMetadataStore].
Required. The resource name of the MetadataStore to retrieve. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Lists MetadataStores for a Location.
Request message for [MetadataService.ListMetadataStores][google.cloud.aiplatform.v1.MetadataService.ListMetadataStores].
Required. The Location whose MetadataStores should be listed. Format: `projects/{project}/locations/{location}`
The maximum number of Metadata Stores to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100.
A page token, received from a previous [MetadataService.ListMetadataStores][google.cloud.aiplatform.v1.MetadataService.ListMetadataStores] call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.)
Response message for [MetadataService.ListMetadataStores][google.cloud.aiplatform.v1.MetadataService.ListMetadataStores].
The MetadataStores found for the Location.
A token, which can be sent as [ListMetadataStoresRequest.page_token][google.cloud.aiplatform.v1.ListMetadataStoresRequest.page_token] to retrieve the next page. If this field is not populated, there are no subsequent pages.
Deletes a single MetadataStore and all its child resources (Artifacts, Executions, and Contexts).
Request message for [MetadataService.DeleteMetadataStore][google.cloud.aiplatform.v1.MetadataService.DeleteMetadataStore].
Required. The resource name of the MetadataStore to delete. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Deprecated: Field is no longer supported.
Creates an Artifact associated with a MetadataStore.
Request message for [MetadataService.CreateArtifact][google.cloud.aiplatform.v1.MetadataService.CreateArtifact].
Required. The resource name of the MetadataStore where the Artifact should be created. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. The Artifact to create.
The {artifact} portion of the resource name with the format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}` If not provided, the Artifact's ID will be a UUID generated by the service. Must be 4-128 characters in length. Valid characters are `/[a-z][0-9]-/`. Must be unique across all Artifacts in the parent MetadataStore. (Otherwise the request will fail with ALREADY_EXISTS, or PERMISSION_DENIED if the caller can't view the preexisting Artifact.)
Retrieves a specific Artifact.
Request message for [MetadataService.GetArtifact][google.cloud.aiplatform.v1.MetadataService.GetArtifact].
Required. The resource name of the Artifact to retrieve. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}`
Lists Artifacts in the MetadataStore.
Request message for [MetadataService.ListArtifacts][google.cloud.aiplatform.v1.MetadataService.ListArtifacts].
Required. The MetadataStore whose Artifacts should be listed. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
The maximum number of Artifacts to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100.
A page token, received from a previous [MetadataService.ListArtifacts][google.cloud.aiplatform.v1.MetadataService.ListArtifacts] call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.)
Filter specifying the boolean condition for the Artifacts to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. The supported set of filters include the following: * **Attribute filtering**: For example: `display_name = "test"`. Supported fields include: `name`, `display_name`, `uri`, `state`, `schema_title`, `create_time`, and `update_time`. Time fields, such as `create_time` and `update_time`, require values specified in RFC-3339 format. For example: `create_time = "2020-11-19T11:30:00-04:00"` * **Metadata field**: To filter on metadata fields use traversal operation as follows: `metadata.<field_name>.<type_value>`. For example: `metadata.field_1.number_value = 10.0` In case the field name contains special characters (such as colon), one can embed it inside double quote. For example: `metadata."field:1".number_value = 10.0` * **Context based filtering**: To filter Artifacts based on the contexts to which they belong, use the function operator with the full resource name `in_context(<context-name>)`. For example: `in_context("projects/<project_number>/locations/<location>/metadataStores/<metadatastore_name>/contexts/<context-id>")` Each of the above supported filter types can be combined together using logical operators (`AND` & `OR`). Maximum nested expression depth allowed is 5. For example: `display_name = "test" AND metadata.field1.bool_value = true`.
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a `.` character, such as foo.bar. see https://google.aip.dev/132#ordering for more details.
Response message for [MetadataService.ListArtifacts][google.cloud.aiplatform.v1.MetadataService.ListArtifacts].
The Artifacts retrieved from the MetadataStore.
A token, which can be sent as [ListArtifactsRequest.page_token][google.cloud.aiplatform.v1.ListArtifactsRequest.page_token] to retrieve the next page. If this field is not populated, there are no subsequent pages.
Updates a stored Artifact.
Request message for [MetadataService.UpdateArtifact][google.cloud.aiplatform.v1.MetadataService.UpdateArtifact].
Required. The Artifact containing updates. The Artifact's [Artifact.name][google.cloud.aiplatform.v1.Artifact.name] field is used to identify the Artifact to be updated. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}`
Optional. A FieldMask indicating which fields should be updated.
If set to true, and the [Artifact][google.cloud.aiplatform.v1.Artifact] is not found, a new [Artifact][google.cloud.aiplatform.v1.Artifact] is created.
Deletes an Artifact.
Request message for [MetadataService.DeleteArtifact][google.cloud.aiplatform.v1.MetadataService.DeleteArtifact].
Required. The resource name of the Artifact to delete. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}`
Optional. The etag of the Artifact to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION.
Purges Artifacts.
Request message for [MetadataService.PurgeArtifacts][google.cloud.aiplatform.v1.MetadataService.PurgeArtifacts].
Required. The metadata store to purge Artifacts from. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. A required filter matching the Artifacts to be purged. E.g., `update_time <= 2020-11-19T11:30:00-04:00`.
Optional. Flag to indicate to actually perform the purge. If `force` is set to false, the method will return a sample of Artifact names that would be deleted.
Creates a Context associated with a MetadataStore.
Request message for [MetadataService.CreateContext][google.cloud.aiplatform.v1.MetadataService.CreateContext].
Required. The resource name of the MetadataStore where the Context should be created. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. The Context to create.
The {context} portion of the resource name with the format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`. If not provided, the Context's ID will be a UUID generated by the service. Must be 4-128 characters in length. Valid characters are `/[a-z][0-9]-/`. Must be unique across all Contexts in the parent MetadataStore. (Otherwise the request will fail with ALREADY_EXISTS, or PERMISSION_DENIED if the caller can't view the preexisting Context.)
Retrieves a specific Context.
Request message for [MetadataService.GetContext][google.cloud.aiplatform.v1.MetadataService.GetContext].
Required. The resource name of the Context to retrieve. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
Lists Contexts on the MetadataStore.
Request message for [MetadataService.ListContexts][google.cloud.aiplatform.v1.MetadataService.ListContexts]
Required. The MetadataStore whose Contexts should be listed. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
The maximum number of Contexts to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100.
A page token, received from a previous [MetadataService.ListContexts][google.cloud.aiplatform.v1.MetadataService.ListContexts] call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.)
Filter specifying the boolean condition for the Contexts to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. Following are the supported set of filters: * **Attribute filtering**: For example: `display_name = "test"`. Supported fields include: `name`, `display_name`, `schema_title`, `create_time`, and `update_time`. Time fields, such as `create_time` and `update_time`, require values specified in RFC-3339 format. For example: `create_time = "2020-11-19T11:30:00-04:00"`. * **Metadata field**: To filter on metadata fields use traversal operation as follows: `metadata.<field_name>.<type_value>`. For example: `metadata.field_1.number_value = 10.0`. In case the field name contains special characters (such as colon), one can embed it inside double quote. For example: `metadata."field:1".number_value = 10.0` * **Parent Child filtering**: To filter Contexts based on parent-child relationship use the HAS operator as follows: ``` parent_contexts: "projects/<project_number>/locations/<location>/metadataStores/<metadatastore_name>/contexts/<context_id>" child_contexts: "projects/<project_number>/locations/<location>/metadataStores/<metadatastore_name>/contexts/<context_id>" ``` Each of the above supported filters can be combined together using logical operators (`AND` & `OR`). Maximum nested expression depth allowed is 5. For example: `display_name = "test" AND metadata.field1.bool_value = true`.
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a `.` character, such as foo.bar. see https://google.aip.dev/132#ordering for more details.
Response message for [MetadataService.ListContexts][google.cloud.aiplatform.v1.MetadataService.ListContexts].
The Contexts retrieved from the MetadataStore.
A token, which can be sent as [ListContextsRequest.page_token][google.cloud.aiplatform.v1.ListContextsRequest.page_token] to retrieve the next page. If this field is not populated, there are no subsequent pages.
Updates a stored Context.
Request message for [MetadataService.UpdateContext][google.cloud.aiplatform.v1.MetadataService.UpdateContext].
Required. The Context containing updates. The Context's [Context.name][google.cloud.aiplatform.v1.Context.name] field is used to identify the Context to be updated. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
Optional. A FieldMask indicating which fields should be updated.
If set to true, and the [Context][google.cloud.aiplatform.v1.Context] is not found, a new [Context][google.cloud.aiplatform.v1.Context] is created.
Deletes a stored Context.
Request message for [MetadataService.DeleteContext][google.cloud.aiplatform.v1.MetadataService.DeleteContext].
Required. The resource name of the Context to delete. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
The force deletion semantics is still undefined. Users should not use this field.
Optional. The etag of the Context to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION.
Purges Contexts.
Request message for [MetadataService.PurgeContexts][google.cloud.aiplatform.v1.MetadataService.PurgeContexts].
Required. The metadata store to purge Contexts from. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. A required filter matching the Contexts to be purged. E.g., `update_time <= 2020-11-19T11:30:00-04:00`.
Optional. Flag to indicate to actually perform the purge. If `force` is set to false, the method will return a sample of Context names that would be deleted.
Adds a set of Artifacts and Executions to a Context. If any of the Artifacts or Executions have already been added to a Context, they are simply skipped.
Request message for [MetadataService.AddContextArtifactsAndExecutions][google.cloud.aiplatform.v1.MetadataService.AddContextArtifactsAndExecutions].
Required. The resource name of the Context that the Artifacts and Executions belong to. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
The resource names of the Artifacts to attribute to the Context. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}`
The resource names of the Executions to associate with the Context. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
Response message for [MetadataService.AddContextArtifactsAndExecutions][google.cloud.aiplatform.v1.MetadataService.AddContextArtifactsAndExecutions].
(message has no fields)
Adds a set of Contexts as children to a parent Context. If any of the child Contexts have already been added to the parent Context, they are simply skipped. If this call would create a cycle or cause any Context to have more than 10 parents, the request will fail with an INVALID_ARGUMENT error.
Request message for [MetadataService.AddContextChildren][google.cloud.aiplatform.v1.MetadataService.AddContextChildren].
Required. The resource name of the parent Context. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
The resource names of the child Contexts.
Response message for [MetadataService.AddContextChildren][google.cloud.aiplatform.v1.MetadataService.AddContextChildren].
(message has no fields)
Remove a set of children contexts from a parent Context. If any of the child Contexts were NOT added to the parent Context, they are simply skipped.
Request message for [MetadataService.DeleteContextChildrenRequest][].
Required. The resource name of the parent Context. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}`
The resource names of the child Contexts.
Response message for [MetadataService.RemoveContextChildren][google.cloud.aiplatform.v1.MetadataService.RemoveContextChildren].
(message has no fields)
Retrieves Artifacts and Executions within the specified Context, connected by Event edges and returned as a LineageSubgraph.
Request message for [MetadataService.QueryContextLineageSubgraph][google.cloud.aiplatform.v1.MetadataService.QueryContextLineageSubgraph].
Required. The resource name of the Context whose Artifacts and Executions should be retrieved as a LineageSubgraph. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/contexts/{context}` The request may error with FAILED_PRECONDITION if the number of Artifacts, the number of Executions, or the number of Events that would be returned for the Context exceeds 1000.
Creates an Execution associated with a MetadataStore.
Request message for [MetadataService.CreateExecution][google.cloud.aiplatform.v1.MetadataService.CreateExecution].
Required. The resource name of the MetadataStore where the Execution should be created. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. The Execution to create.
The {execution} portion of the resource name with the format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}` If not provided, the Execution's ID will be a UUID generated by the service. Must be 4-128 characters in length. Valid characters are `/[a-z][0-9]-/`. Must be unique across all Executions in the parent MetadataStore. (Otherwise the request will fail with ALREADY_EXISTS, or PERMISSION_DENIED if the caller can't view the preexisting Execution.)
Retrieves a specific Execution.
Request message for [MetadataService.GetExecution][google.cloud.aiplatform.v1.MetadataService.GetExecution].
Required. The resource name of the Execution to retrieve. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
Lists Executions in the MetadataStore.
Request message for [MetadataService.ListExecutions][google.cloud.aiplatform.v1.MetadataService.ListExecutions].
Required. The MetadataStore whose Executions should be listed. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
The maximum number of Executions to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100.
A page token, received from a previous [MetadataService.ListExecutions][google.cloud.aiplatform.v1.MetadataService.ListExecutions] call. Provide this to retrieve the subsequent page. When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with an INVALID_ARGUMENT error.)
Filter specifying the boolean condition for the Executions to satisfy in order to be part of the result set. The syntax to define filter query is based on https://google.aip.dev/160. Following are the supported set of filters: * **Attribute filtering**: For example: `display_name = "test"`. Supported fields include: `name`, `display_name`, `state`, `schema_title`, `create_time`, and `update_time`. Time fields, such as `create_time` and `update_time`, require values specified in RFC-3339 format. For example: `create_time = "2020-11-19T11:30:00-04:00"`. * **Metadata field**: To filter on metadata fields use traversal operation as follows: `metadata.<field_name>.<type_value>` For example: `metadata.field_1.number_value = 10.0` In case the field name contains special characters (such as colon), one can embed it inside double quote. For example: `metadata."field:1".number_value = 10.0` * **Context based filtering**: To filter Executions based on the contexts to which they belong use the function operator with the full resource name: `in_context(<context-name>)`. For example: `in_context("projects/<project_number>/locations/<location>/metadataStores/<metadatastore_name>/contexts/<context-id>")` Each of the above supported filters can be combined together using logical operators (`AND` & `OR`). Maximum nested expression depth allowed is 5. For example: `display_name = "test" AND metadata.field1.bool_value = true`.
How the list of messages is ordered. Specify the values to order by and an ordering operation. The default sorting order is ascending. To specify descending order for a field, users append a " desc" suffix; for example: "foo desc, bar". Subfields are specified with a `.` character, such as foo.bar. see https://google.aip.dev/132#ordering for more details.
Response message for [MetadataService.ListExecutions][google.cloud.aiplatform.v1.MetadataService.ListExecutions].
The Executions retrieved from the MetadataStore.
A token, which can be sent as [ListExecutionsRequest.page_token][google.cloud.aiplatform.v1.ListExecutionsRequest.page_token] to retrieve the next page. If this field is not populated, there are no subsequent pages.
Updates a stored Execution.
Request message for [MetadataService.UpdateExecution][google.cloud.aiplatform.v1.MetadataService.UpdateExecution].
Required. The Execution containing updates. The Execution's [Execution.name][google.cloud.aiplatform.v1.Execution.name] field is used to identify the Execution to be updated. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
Optional. A FieldMask indicating which fields should be updated.
If set to true, and the [Execution][google.cloud.aiplatform.v1.Execution] is not found, a new [Execution][google.cloud.aiplatform.v1.Execution] is created.
Deletes an Execution.
Request message for [MetadataService.DeleteExecution][google.cloud.aiplatform.v1.MetadataService.DeleteExecution].
Required. The resource name of the Execution to delete. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
Optional. The etag of the Execution to delete. If this is provided, it must match the server's etag. Otherwise, the request will fail with a FAILED_PRECONDITION.
Purges Executions.
Request message for [MetadataService.PurgeExecutions][google.cloud.aiplatform.v1.MetadataService.PurgeExecutions].
Required. The metadata store to purge Executions from. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. A required filter matching the Executions to be purged. E.g., `update_time <= 2020-11-19T11:30:00-04:00`.
Optional. Flag to indicate to actually perform the purge. If `force` is set to false, the method will return a sample of Execution names that would be deleted.
Adds Events to the specified Execution. An Event indicates whether an Artifact was used as an input or output for an Execution. If an Event already exists between the Execution and the Artifact, the Event is skipped.
Request message for [MetadataService.AddExecutionEvents][google.cloud.aiplatform.v1.MetadataService.AddExecutionEvents].
Required. The resource name of the Execution that the Events connect Artifacts with. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
The Events to create and add.
Response message for [MetadataService.AddExecutionEvents][google.cloud.aiplatform.v1.MetadataService.AddExecutionEvents].
(message has no fields)
Obtains the set of input and output Artifacts for this Execution, in the form of LineageSubgraph that also contains the Execution and connecting Events.
Request message for [MetadataService.QueryExecutionInputsAndOutputs][google.cloud.aiplatform.v1.MetadataService.QueryExecutionInputsAndOutputs].
Required. The resource name of the Execution whose input and output Artifacts should be retrieved as a LineageSubgraph. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/executions/{execution}`
Creates a MetadataSchema.
Request message for [MetadataService.CreateMetadataSchema][google.cloud.aiplatform.v1.MetadataService.CreateMetadataSchema].
Required. The resource name of the MetadataStore where the MetadataSchema should be created. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
Required. The MetadataSchema to create.
The {metadata_schema} portion of the resource name with the format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/metadataSchemas/{metadataschema}` If not provided, the MetadataStore's ID will be a UUID generated by the service. Must be 4-128 characters in length. Valid characters are `/[a-z][0-9]-/`. Must be unique across all MetadataSchemas in the parent Location. (Otherwise the request will fail with ALREADY_EXISTS, or PERMISSION_DENIED if the caller can't view the preexisting MetadataSchema.)
Retrieves a specific MetadataSchema.
Request message for [MetadataService.GetMetadataSchema][google.cloud.aiplatform.v1.MetadataService.GetMetadataSchema].
Required. The resource name of the MetadataSchema to retrieve. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/metadataSchemas/{metadataschema}`
Lists MetadataSchemas.
Request message for [MetadataService.ListMetadataSchemas][google.cloud.aiplatform.v1.MetadataService.ListMetadataSchemas].
Required. The MetadataStore whose MetadataSchemas should be listed. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}`
The maximum number of MetadataSchemas to return. The service may return fewer. Must be in range 1-1000, inclusive. Defaults to 100.
A page token, received from a previous [MetadataService.ListMetadataSchemas][google.cloud.aiplatform.v1.MetadataService.ListMetadataSchemas] call. Provide this to retrieve the next page. When paginating, all other provided parameters must match the call that provided the page token. (Otherwise the request will fail with INVALID_ARGUMENT error.)
A query to filter available MetadataSchemas for matching results.
Response message for [MetadataService.ListMetadataSchemas][google.cloud.aiplatform.v1.MetadataService.ListMetadataSchemas].
The MetadataSchemas found for the MetadataStore.
A token, which can be sent as [ListMetadataSchemasRequest.page_token][google.cloud.aiplatform.v1.ListMetadataSchemasRequest.page_token] to retrieve the next page. If this field is not populated, there are no subsequent pages.
Retrieves lineage of an Artifact represented through Artifacts and Executions connected by Event edges and returned as a LineageSubgraph.
Request message for [MetadataService.QueryArtifactLineageSubgraph][google.cloud.aiplatform.v1.MetadataService.QueryArtifactLineageSubgraph].
Required. The resource name of the Artifact whose Lineage needs to be retrieved as a LineageSubgraph. Format: `projects/{project}/locations/{location}/metadataStores/{metadatastore}/artifacts/{artifact}` The request may error with FAILED_PRECONDITION if the number of Artifacts, the number of Executions, or the number of Events that would be returned for the Context exceeds 1000.
Specifies the size of the lineage graph in terms of number of hops from the specified artifact. Negative Value: INVALID_ARGUMENT error is returned 0: Only input artifact is returned. No value: Transitive closure is performed to return the complete graph.
Filter specifying the boolean condition for the Artifacts to satisfy in order to be part of the Lineage Subgraph. The syntax to define filter query is based on https://google.aip.dev/160. The supported set of filters include the following: * **Attribute filtering**: For example: `display_name = "test"` Supported fields include: `name`, `display_name`, `uri`, `state`, `schema_title`, `create_time`, and `update_time`. Time fields, such as `create_time` and `update_time`, require values specified in RFC-3339 format. For example: `create_time = "2020-11-19T11:30:00-04:00"` * **Metadata field**: To filter on metadata fields use traversal operation as follows: `metadata.<field_name>.<type_value>`. For example: `metadata.field_1.number_value = 10.0` In case the field name contains special characters (such as colon), one can embed it inside double quote. For example: `metadata."field:1".number_value = 10.0` Each of the above supported filter types can be combined together using logical operators (`AND` & `OR`). Maximum nested expression depth allowed is 5. For example: `display_name = "test" AND metadata.field1.bool_value = true`.
A service that migrates resources from automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com to Vertex AI.
Searches all of the resources in automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com that can be migrated to Vertex AI's given location.
Request message for [MigrationService.SearchMigratableResources][google.cloud.aiplatform.v1.MigrationService.SearchMigratableResources].
Required. The location that the migratable resources should be searched from. It's the Vertex AI location that the resources can be migrated to, not the resources' original location. Format: `projects/{project}/locations/{location}`
The standard page size. The default and maximum value is 100.
The standard page token.
A filter for your search. You can use the following types of filters: * Resource type filters. The following strings filter for a specific type of [MigratableResource][google.cloud.aiplatform.v1.MigratableResource]: * `ml_engine_model_version:*` * `automl_model:*` * `automl_dataset:*` * `data_labeling_dataset:*` * "Migrated or not" filters. The following strings filter for resources that either have or have not already been migrated: * `last_migrate_time:*` filters for migrated resources. * `NOT last_migrate_time:*` filters for not yet migrated resources.
Response message for [MigrationService.SearchMigratableResources][google.cloud.aiplatform.v1.MigrationService.SearchMigratableResources].
All migratable resources that can be migrated to the location specified in the request.
The standard next-page token. The migratable_resources may not fill page_size in SearchMigratableResourcesRequest even when there are subsequent pages.
Batch migrates resources from ml.googleapis.com, automl.googleapis.com, and datalabeling.googleapis.com to Vertex AI.
Request message for [MigrationService.BatchMigrateResources][google.cloud.aiplatform.v1.MigrationService.BatchMigrateResources].
Required. The location of the migrated resource will live in. Format: `projects/{project}/locations/{location}`
Required. The request messages specifying the resources to migrate. They must be in the same location as the destination. Up to 50 resources can be migrated in one batch.
The interface of Model Garden Service.
Gets a Model Garden publisher model.
Request message for [ModelGardenService.GetPublisherModel][google.cloud.aiplatform.v1.ModelGardenService.GetPublisherModel]
Required. The name of the PublisherModel resource. Format: `publishers/{publisher}/models/{publisher_model}`
Optional. The IETF BCP-47 language code representing the language in which the publisher model's text information should be written in.
Optional. PublisherModel view specifying which fields to read.
Optional. Boolean indicates whether the requested model is a Hugging Face model.
Optional. Token used to access Hugging Face gated models.
A Model Garden Publisher Model.
Output only. The resource name of the PublisherModel.
Output only. Immutable. The version ID of the PublisherModel. A new version is committed when a new model version is uploaded under an existing model id. It is an auto-incrementing decimal number in string representation.
Required. Indicates the open source category of the publisher model.
Optional. Supported call-to-action options.
Optional. Additional information about the model's Frameworks.
Optional. Indicates the launch stage of the model.
Optional. Indicates the state of the model version.
Optional. Output only. Immutable. Used to indicate this model has a publisher model and provide the template of the publisher model resource name.
Optional. The schemata that describes formats of the PublisherModel's predictions and explanations as given and returned via [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict].
A service for managing Vertex AI's machine learning Models.
Uploads a Model artifact into Vertex AI.
Request message for [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel].
Required. The resource name of the Location into which to upload the Model. Format: `projects/{project}/locations/{location}`
Optional. The resource name of the model into which to upload the version. Only specify this field when uploading a new version.
Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are `[a-z0-9_-]`. The first character cannot be a number or hyphen.
Required. The Model to create.
Optional. The user-provided custom service account to use to do the model upload. If empty, [Vertex AI Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) will be used to access resources needed to upload the model. This account must belong to the target project where the model is uploaded to, i.e., the project specified in the `parent` field of this request and have necessary read permissions (to Google Cloud Storage, Artifact Registry, etc.).
Gets a Model.
Request message for [ModelService.GetModel][google.cloud.aiplatform.v1.ModelService.GetModel].
Required. The name of the Model resource. Format: `projects/{project}/locations/{location}/models/{model}` In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` If no version ID or alias is specified, the "default" version will be returned. The "default" version alias is created for the first version of the model, and can be moved to other versions later on. There will be exactly one default version.
Lists Models in a Location.
Request message for [ModelService.ListModels][google.cloud.aiplatform.v1.ModelService.ListModels].
Required. The resource name of the Location to list the Models from. Format: `projects/{project}/locations/{location}`
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `model` supports = and !=. `model` represents the Model ID, i.e. the last segment of the Model's [resource name][google.cloud.aiplatform.v1.Model.name]. * `display_name` supports = and != * `labels` supports general map functions that is: * `labels.key=value` - key:value equality * `labels.key:* or labels:key - key existence * A key including a space must be quoted. `labels."a key"`. * `base_model_name` only supports = Some examples: * `model=1234` * `displayName="myDisplayName"` * `labels.myKey="myValue"` * `baseModelName="text-bison"`
The standard list page size.
The standard list page token. Typically obtained via [ListModelsResponse.next_page_token][google.cloud.aiplatform.v1.ListModelsResponse.next_page_token] of the previous [ModelService.ListModels][google.cloud.aiplatform.v1.ModelService.ListModels] call.
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time` Example: `display_name, create_time desc`.
Response message for [ModelService.ListModels][google.cloud.aiplatform.v1.ModelService.ListModels]
List of Models in the requested page.
A token to retrieve next page of results. Pass to [ListModelsRequest.page_token][google.cloud.aiplatform.v1.ListModelsRequest.page_token] to obtain that page.
Lists versions of the specified model.
Request message for [ModelService.ListModelVersions][google.cloud.aiplatform.v1.ModelService.ListModelVersions].
Required. The name of the model to list versions for.
The standard list page size.
The standard list page token. Typically obtained via [next_page_token][google.cloud.aiplatform.v1.ListModelVersionsResponse.next_page_token] of the previous [ListModelVersions][google.cloud.aiplatform.v1.ModelService.ListModelVersions] call.
An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `labels` supports general map functions that is: * `labels.key=value` - key:value equality * `labels.key:* or labels:key - key existence * A key including a space must be quoted. `labels."a key"`. Some examples: * `labels.myKey="myValue"`
Mask specifying which fields to read.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `create_time` * `update_time` Example: `update_time asc, create_time desc`.
Response message for [ModelService.ListModelVersions][google.cloud.aiplatform.v1.ModelService.ListModelVersions]
List of Model versions in the requested page. In the returned Model name field, version ID instead of regvision tag will be included.
A token to retrieve the next page of results. Pass to [ListModelVersionsRequest.page_token][google.cloud.aiplatform.v1.ListModelVersionsRequest.page_token] to obtain that page.
Lists checkpoints of the specified model version.
Request message for [ModelService.ListModelVersionCheckpoints][google.cloud.aiplatform.v1.ModelService.ListModelVersionCheckpoints].
Required. The name of the model version to list checkpoints for. `projects/{project}/locations/{location}/models/{model}@{version}` Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` If no version ID or alias is specified, the latest version will be used.
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [next_page_token][google.cloud.aiplatform.v1.ListModelVersionCheckpointsResponse.next_page_token] of the previous [ListModelVersionCheckpoints][google.cloud.aiplatform.v1.ModelService.ListModelVersionCheckpoints] call.
Response message for [ModelService.ListModelVersionCheckpoints][google.cloud.aiplatform.v1.ModelService.ListModelVersionCheckpoints]
List of Model Version checkpoints.
A token to retrieve the next page of results. Pass to [ListModelVersionCheckpointsRequest.page_token][google.cloud.aiplatform.v1.ListModelVersionCheckpointsRequest.page_token] to obtain that page.
Updates a Model.
Request message for [ModelService.UpdateModel][google.cloud.aiplatform.v1.ModelService.UpdateModel].
Required. The Model which replaces the resource on the server. When Model Versioning is enabled, the model.name will be used to determine whether to update the model or model version. 1. model.name with the @ value, e.g. models/123@1, refers to a version specific update. 2. model.name without the @ value, e.g. models/123, refers to a model update. 3. model.name with @-, e.g. models/123@-, refers to a model update. 4. Supported model fields: display_name, description; supported version-specific fields: version_description. Labels are supported in both scenarios. Both the model labels and the version labels are merged when a model is returned. When updating labels, if the request is for model-specific update, model label gets updated. Otherwise, version labels get updated. 5. A model name or model version name fields update mismatch will cause a precondition error. 6. One request cannot update both the model and the version fields. You must update them separately.
Required. The update mask applies to the resource. For the `FieldMask` definition, see [google.protobuf.FieldMask][google.protobuf.FieldMask].
Incrementally update the dataset used for an examples model.
Request message for [ModelService.UpdateExplanationDataset][google.cloud.aiplatform.v1.ModelService.UpdateExplanationDataset].
Required. The resource name of the Model to update. Format: `projects/{project}/locations/{location}/models/{model}`
The example config containing the location of the dataset.
Deletes a Model. A model cannot be deleted if any [Endpoint][google.cloud.aiplatform.v1.Endpoint] resource has a [DeployedModel][google.cloud.aiplatform.v1.DeployedModel] based on the model in its [deployed_models][google.cloud.aiplatform.v1.Endpoint.deployed_models] field.
Request message for [ModelService.DeleteModel][google.cloud.aiplatform.v1.ModelService.DeleteModel].
Required. The name of the Model resource to be deleted. Format: `projects/{project}/locations/{location}/models/{model}`
Deletes a Model version. Model version can only be deleted if there are no [DeployedModels][google.cloud.aiplatform.v1.DeployedModel] created from it. Deleting the only version in the Model is not allowed. Use [DeleteModel][google.cloud.aiplatform.v1.ModelService.DeleteModel] for deleting the Model instead.
Request message for [ModelService.DeleteModelVersion][google.cloud.aiplatform.v1.ModelService.DeleteModelVersion].
Required. The name of the model version to be deleted, with a version ID explicitly included. Example: `projects/{project}/locations/{location}/models/{model}@1234`
Merges a set of aliases for a Model version.
Request message for [ModelService.MergeVersionAliases][google.cloud.aiplatform.v1.ModelService.MergeVersionAliases].
Required. The name of the model version to merge aliases, with a version ID explicitly included. Example: `projects/{project}/locations/{location}/models/{model}@1234`
Required. The set of version aliases to merge. The alias should be at most 128 characters, and match `[a-z][a-zA-Z0-9-]{0,126}[a-z-0-9]`. Add the `-` prefix to an alias means removing that alias from the version. `-` is NOT counted in the 128 characters. Example: `-golden` means removing the `golden` alias from the version. There is NO ordering in aliases, which means 1) The aliases returned from GetModel API might not have the exactly same order from this MergeVersionAliases API. 2) Adding and deleting the same alias in the request is not recommended, and the 2 operations will be cancelled out.
Exports a trained, exportable Model to a location specified by the user. A Model is considered to be exportable if it has at least one [supported export format][google.cloud.aiplatform.v1.Model.supported_export_formats].
Request message for [ModelService.ExportModel][google.cloud.aiplatform.v1.ModelService.ExportModel].
Required. The resource name of the Model to export. The resource name may contain version id or version alias to specify the version, if no version is specified, the default version will be exported.
Required. The desired output location and configuration.
Copies an already existing Vertex AI Model into the specified Location. The source Model must exist in the same Project. When copying custom Models, the users themselves are responsible for [Model.metadata][google.cloud.aiplatform.v1.Model.metadata] content to be region-agnostic, as well as making sure that any resources (e.g. files) it depends on remain accessible.
Request message for [ModelService.CopyModel][google.cloud.aiplatform.v1.ModelService.CopyModel].
If both fields are unset, a new Model will be created with a generated ID.
Optional. Copy source_model into a new Model with this ID. The ID will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are `[a-z0-9_-]`. The first character cannot be a number or hyphen.
Optional. Specify this field to copy source_model into this existing Model as a new version. Format: `projects/{project}/locations/{location}/models/{model}`
Required. The resource name of the Location into which to copy the Model. Format: `projects/{project}/locations/{location}`
Required. The resource name of the Model to copy. That Model must be in the same Project. Format: `projects/{project}/locations/{location}/models/{model}`
Customer-managed encryption key options. If this is set, then the Model copy will be encrypted with the provided encryption key.
Imports an externally generated ModelEvaluation.
Request message for [ModelService.ImportModelEvaluation][google.cloud.aiplatform.v1.ModelService.ImportModelEvaluation]
Required. The name of the parent model resource. Format: `projects/{project}/locations/{location}/models/{model}`
Required. Model evaluation resource to be imported.
Imports a list of externally generated ModelEvaluationSlice.
Request message for [ModelService.BatchImportModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.BatchImportModelEvaluationSlices]
Required. The name of the parent ModelEvaluation resource. Format: `projects/{project}/locations/{location}/models/{model}/evaluations/{evaluation}`
Required. Model evaluation slice resource to be imported.
Response message for [ModelService.BatchImportModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.BatchImportModelEvaluationSlices]
Output only. List of imported [ModelEvaluationSlice.name][google.cloud.aiplatform.v1.ModelEvaluationSlice.name].
Imports a list of externally generated EvaluatedAnnotations.
Request message for [ModelService.BatchImportEvaluatedAnnotations][google.cloud.aiplatform.v1.ModelService.BatchImportEvaluatedAnnotations]
Required. The name of the parent ModelEvaluationSlice resource. Format: `projects/{project}/locations/{location}/models/{model}/evaluations/{evaluation}/slices/{slice}`
Required. Evaluated annotations resource to be imported.
Response message for [ModelService.BatchImportEvaluatedAnnotations][google.cloud.aiplatform.v1.ModelService.BatchImportEvaluatedAnnotations]
Output only. Number of EvaluatedAnnotations imported.
Gets a ModelEvaluation.
Request message for [ModelService.GetModelEvaluation][google.cloud.aiplatform.v1.ModelService.GetModelEvaluation].
Required. The name of the ModelEvaluation resource. Format: `projects/{project}/locations/{location}/models/{model}/evaluations/{evaluation}`
Lists ModelEvaluations in a Model.
Request message for [ModelService.ListModelEvaluations][google.cloud.aiplatform.v1.ModelService.ListModelEvaluations].
Required. The resource name of the Model to list the ModelEvaluations from. Format: `projects/{project}/locations/{location}/models/{model}`
The standard list filter.
The standard list page size.
The standard list page token. Typically obtained via [ListModelEvaluationsResponse.next_page_token][google.cloud.aiplatform.v1.ListModelEvaluationsResponse.next_page_token] of the previous [ModelService.ListModelEvaluations][google.cloud.aiplatform.v1.ModelService.ListModelEvaluations] call.
Mask specifying which fields to read.
Response message for [ModelService.ListModelEvaluations][google.cloud.aiplatform.v1.ModelService.ListModelEvaluations].
List of ModelEvaluations in the requested page.
A token to retrieve next page of results. Pass to [ListModelEvaluationsRequest.page_token][google.cloud.aiplatform.v1.ListModelEvaluationsRequest.page_token] to obtain that page.
Gets a ModelEvaluationSlice.
Request message for [ModelService.GetModelEvaluationSlice][google.cloud.aiplatform.v1.ModelService.GetModelEvaluationSlice].
Required. The name of the ModelEvaluationSlice resource. Format: `projects/{project}/locations/{location}/models/{model}/evaluations/{evaluation}/slices/{slice}`
Lists ModelEvaluationSlices in a ModelEvaluation.
Request message for [ModelService.ListModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.ListModelEvaluationSlices].
Required. The resource name of the ModelEvaluation to list the ModelEvaluationSlices from. Format: `projects/{project}/locations/{location}/models/{model}/evaluations/{evaluation}`
The standard list filter. * `slice.dimension` - for =.
The standard list page size.
The standard list page token. Typically obtained via [ListModelEvaluationSlicesResponse.next_page_token][google.cloud.aiplatform.v1.ListModelEvaluationSlicesResponse.next_page_token] of the previous [ModelService.ListModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.ListModelEvaluationSlices] call.
Mask specifying which fields to read.
Response message for [ModelService.ListModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.ListModelEvaluationSlices].
List of ModelEvaluations in the requested page.
A token to retrieve next page of results. Pass to [ListModelEvaluationSlicesRequest.page_token][google.cloud.aiplatform.v1.ListModelEvaluationSlicesRequest.page_token] to obtain that page.
The interface for Vertex Notebook service (a.k.a. Colab on Workbench).
Creates a NotebookRuntimeTemplate.
Request message for [NotebookService.CreateNotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookService.CreateNotebookRuntimeTemplate].
Required. The resource name of the Location to create the NotebookRuntimeTemplate. Format: `projects/{project}/locations/{location}`
Required. The NotebookRuntimeTemplate to create.
Optional. User specified ID for the notebook runtime template.
Gets a NotebookRuntimeTemplate.
Request message for [NotebookService.GetNotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookService.GetNotebookRuntimeTemplate]
Required. The name of the NotebookRuntimeTemplate resource. Format: `projects/{project}/locations/{location}/notebookRuntimeTemplates/{notebook_runtime_template}`
Lists NotebookRuntimeTemplates in a Location.
Request message for [NotebookService.ListNotebookRuntimeTemplates][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimeTemplates].
Required. The resource name of the Location from which to list the NotebookRuntimeTemplates. Format: `projects/{project}/locations/{location}`
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `notebookRuntimeTemplate` supports = and !=. `notebookRuntimeTemplate` represents the NotebookRuntimeTemplate ID, i.e. the last segment of the NotebookRuntimeTemplate's [resource name] [google.cloud.aiplatform.v1.NotebookRuntimeTemplate.name]. * `display_name` supports = and != * `labels` supports general map functions that is: * `labels.key=value` - key:value equality * `labels.key:* or labels:key - key existence * A key including a space must be quoted. `labels."a key"`. * `notebookRuntimeType` supports = and !=. notebookRuntimeType enum: [USER_DEFINED, ONE_CLICK]. * `machineType` supports = and !=. * `acceleratorType` supports = and !=. Some examples: * `notebookRuntimeTemplate=notebookRuntimeTemplate123` * `displayName="myDisplayName"` * `labels.myKey="myValue"` * `notebookRuntimeType=USER_DEFINED` * `machineType=e2-standard-4` * `acceleratorType=NVIDIA_TESLA_T4`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListNotebookRuntimeTemplatesResponse.next_page_token][google.cloud.aiplatform.v1.ListNotebookRuntimeTemplatesResponse.next_page_token] of the previous [NotebookService.ListNotebookRuntimeTemplates][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimeTemplates] call.
Optional. Mask specifying which fields to read.
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time` Example: `display_name, create_time desc`.
Response message for [NotebookService.ListNotebookRuntimeTemplates][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimeTemplates].
List of NotebookRuntimeTemplates in the requested page.
A token to retrieve next page of results. Pass to [ListNotebookRuntimeTemplatesRequest.page_token][google.cloud.aiplatform.v1.ListNotebookRuntimeTemplatesRequest.page_token] to obtain that page.
Deletes a NotebookRuntimeTemplate.
Request message for [NotebookService.DeleteNotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookService.DeleteNotebookRuntimeTemplate].
Required. The name of the NotebookRuntimeTemplate resource to be deleted. Format: `projects/{project}/locations/{location}/notebookRuntimeTemplates/{notebook_runtime_template}`
Updates a NotebookRuntimeTemplate.
Request message for [NotebookService.UpdateNotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookService.UpdateNotebookRuntimeTemplate].
Required. The NotebookRuntimeTemplate to update.
Required. The update mask applies to the resource. For the `FieldMask` definition, see [google.protobuf.FieldMask][google.protobuf.FieldMask]. Input format: `{paths: "${updated_filed}"}` Updatable fields: * `encryption_spec.kms_key_name`
Assigns a NotebookRuntime to a user for a particular Notebook file. This method will either returns an existing assignment or generates a new one.
Request message for [NotebookService.AssignNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.AssignNotebookRuntime].
Required. The resource name of the Location to get the NotebookRuntime assignment. Format: `projects/{project}/locations/{location}`
Required. The resource name of the NotebookRuntimeTemplate based on which a NotebookRuntime will be assigned (reuse or create a new one).
Required. Provide runtime specific information (e.g. runtime owner, notebook id) used for NotebookRuntime assignment.
Optional. User specified ID for the notebook runtime.
Gets a NotebookRuntime.
Request message for [NotebookService.GetNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.GetNotebookRuntime]
Required. The name of the NotebookRuntime resource. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner.
Lists NotebookRuntimes in a Location.
Request message for [NotebookService.ListNotebookRuntimes][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimes].
Required. The resource name of the Location from which to list the NotebookRuntimes. Format: `projects/{project}/locations/{location}`
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `notebookRuntime` supports = and !=. `notebookRuntime` represents the NotebookRuntime ID, i.e. the last segment of the NotebookRuntime's [resource name] [google.cloud.aiplatform.v1.NotebookRuntime.name]. * `displayName` supports = and != and regex. * `notebookRuntimeTemplate` supports = and !=. `notebookRuntimeTemplate` represents the NotebookRuntimeTemplate ID, i.e. the last segment of the NotebookRuntimeTemplate's [resource name] [google.cloud.aiplatform.v1.NotebookRuntimeTemplate.name]. * `healthState` supports = and !=. healthState enum: [HEALTHY, UNHEALTHY, HEALTH_STATE_UNSPECIFIED]. * `runtimeState` supports = and !=. runtimeState enum: [RUNTIME_STATE_UNSPECIFIED, RUNNING, BEING_STARTED, BEING_STOPPED, STOPPED, BEING_UPGRADED, ERROR, INVALID]. * `runtimeUser` supports = and !=. * API version is UI only: `uiState` supports = and !=. uiState enum: [UI_RESOURCE_STATE_UNSPECIFIED, UI_RESOURCE_STATE_BEING_CREATED, UI_RESOURCE_STATE_ACTIVE, UI_RESOURCE_STATE_BEING_DELETED, UI_RESOURCE_STATE_CREATION_FAILED]. * `notebookRuntimeType` supports = and !=. notebookRuntimeType enum: [USER_DEFINED, ONE_CLICK]. * `machineType` supports = and !=. * `acceleratorType` supports = and !=. Some examples: * `notebookRuntime="notebookRuntime123"` * `displayName="myDisplayName"` and `displayName=~"myDisplayNameRegex"` * `notebookRuntimeTemplate="notebookRuntimeTemplate321"` * `healthState=HEALTHY` * `runtimeState=RUNNING` * `runtimeUser="test@google.com"` * `uiState=UI_RESOURCE_STATE_BEING_DELETED` * `notebookRuntimeType=USER_DEFINED` * `machineType=e2-standard-4` * `acceleratorType=NVIDIA_TESLA_T4`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListNotebookRuntimesResponse.next_page_token][google.cloud.aiplatform.v1.ListNotebookRuntimesResponse.next_page_token] of the previous [NotebookService.ListNotebookRuntimes][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimes] call.
Optional. Mask specifying which fields to read.
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time` Example: `display_name, create_time desc`.
Response message for [NotebookService.ListNotebookRuntimes][google.cloud.aiplatform.v1.NotebookService.ListNotebookRuntimes].
List of NotebookRuntimes in the requested page.
A token to retrieve next page of results. Pass to [ListNotebookRuntimesRequest.page_token][google.cloud.aiplatform.v1.ListNotebookRuntimesRequest.page_token] to obtain that page.
Deletes a NotebookRuntime.
Request message for [NotebookService.DeleteNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.DeleteNotebookRuntime].
Required. The name of the NotebookRuntime resource to be deleted. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner.
Upgrades a NotebookRuntime.
Request message for [NotebookService.UpgradeNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.UpgradeNotebookRuntime].
Required. The name of the NotebookRuntime resource to be upgrade. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner.
Starts a NotebookRuntime.
Request message for [NotebookService.StartNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StartNotebookRuntime].
Required. The name of the NotebookRuntime resource to be started. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner.
Stops a NotebookRuntime.
Request message for [NotebookService.StopNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StopNotebookRuntime].
Required. The name of the NotebookRuntime resource to be stopped. Instead of checking whether the name is in valid NotebookRuntime resource name format, directly throw NotFound exception if there is no such NotebookRuntime in spanner.
Creates a NotebookExecutionJob.
Gets a NotebookExecutionJob.
Request message for [NotebookService.GetNotebookExecutionJob]
Required. The name of the NotebookExecutionJob resource.
Optional. The NotebookExecutionJob view. Defaults to BASIC.
Lists NotebookExecutionJobs in a Location.
Request message for [NotebookService.ListNotebookExecutionJobs]
Required. The resource name of the Location from which to list the NotebookExecutionJobs. Format: `projects/{project}/locations/{location}`
Optional. An expression for filtering the results of the request. For field names both snake_case and camelCase are supported. * `notebookExecutionJob` supports = and !=. `notebookExecutionJob` represents the NotebookExecutionJob ID. * `displayName` supports = and != and regex. * `schedule` supports = and != and regex. Some examples: * `notebookExecutionJob="123"` * `notebookExecutionJob="my-execution-job"` * `displayName="myDisplayName"` and `displayName=~"myDisplayNameRegex"`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListNotebookExecutionJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListNotebookExecutionJobsResponse.next_page_token] of the previous [NotebookService.ListNotebookExecutionJobs][google.cloud.aiplatform.v1.NotebookService.ListNotebookExecutionJobs] call.
Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `display_name` * `create_time` * `update_time` Example: `display_name, create_time desc`.
Optional. The NotebookExecutionJob view. Defaults to BASIC.
Response message for [NotebookService.CreateNotebookExecutionJob]
List of NotebookExecutionJobs in the requested page.
A token to retrieve next page of results. Pass to [ListNotebookExecutionJobsRequest.page_token][google.cloud.aiplatform.v1.ListNotebookExecutionJobsRequest.page_token] to obtain that page.
Deletes a NotebookExecutionJob.
Request message for [NotebookService.DeleteNotebookExecutionJob]
Required. The name of the NotebookExecutionJob resource to be deleted.
A service for managing Vertex AI's machine learning PersistentResource.
Creates a PersistentResource.
Request message for [PersistentResourceService.CreatePersistentResource][google.cloud.aiplatform.v1.PersistentResourceService.CreatePersistentResource].
Required. The resource name of the Location to create the PersistentResource in. Format: `projects/{project}/locations/{location}`
Required. The PersistentResource to create.
Required. The ID to use for the PersistentResource, which become the final component of the PersistentResource's resource name. The maximum length is 63 characters, and valid characters are `/^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$/`.
Gets a PersistentResource.
Request message for [PersistentResourceService.GetPersistentResource][google.cloud.aiplatform.v1.PersistentResourceService.GetPersistentResource].
Required. The name of the PersistentResource resource. Format: `projects/{project_id_or_number}/locations/{location_id}/persistentResources/{persistent_resource_id}`
Lists PersistentResources in a Location.
Request message for [PersistentResourceService.ListPersistentResources][google.cloud.aiplatform.v1.PersistentResourceService.ListPersistentResources].
Required. The resource name of the Location to list the PersistentResources from. Format: `projects/{project}/locations/{location}`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListPersistentResourcesResponse.next_page_token][google.cloud.aiplatform.v1.ListPersistentResourcesResponse.next_page_token] of the previous [PersistentResourceService.ListPersistentResource][] call.
Response message for [PersistentResourceService.ListPersistentResources][google.cloud.aiplatform.v1.PersistentResourceService.ListPersistentResources]
A token to retrieve next page of results. Pass to [ListPersistentResourcesRequest.page_token][google.cloud.aiplatform.v1.ListPersistentResourcesRequest.page_token] to obtain that page.
Deletes a PersistentResource.
Request message for [PersistentResourceService.DeletePersistentResource][google.cloud.aiplatform.v1.PersistentResourceService.DeletePersistentResource].
Required. The name of the PersistentResource to be deleted. Format: `projects/{project}/locations/{location}/persistentResources/{persistent_resource}`
Updates a PersistentResource.
Request message for UpdatePersistentResource method.
Required. The PersistentResource to update. The PersistentResource's `name` field is used to identify the PersistentResource to update. Format: `projects/{project}/locations/{location}/persistentResources/{persistent_resource}`
Required. Specify the fields to be overwritten in the PersistentResource by the update method.
Reboots a PersistentResource.
Request message for [PersistentResourceService.RebootPersistentResource][google.cloud.aiplatform.v1.PersistentResourceService.RebootPersistentResource].
Required. The name of the PersistentResource resource. Format: `projects/{project_id_or_number}/locations/{location_id}/persistentResources/{persistent_resource_id}`
A service for creating and managing Vertex AI's pipelines. This includes both `TrainingPipeline` resources (used for AutoML and custom training) and `PipelineJob` resources (used for Vertex AI Pipelines).
Creates a TrainingPipeline. A created TrainingPipeline right away will be attempted to be run.
Request message for [PipelineService.CreateTrainingPipeline][google.cloud.aiplatform.v1.PipelineService.CreateTrainingPipeline].
Required. The resource name of the Location to create the TrainingPipeline in. Format: `projects/{project}/locations/{location}`
Required. The TrainingPipeline to create.
Gets a TrainingPipeline.
Request message for [PipelineService.GetTrainingPipeline][google.cloud.aiplatform.v1.PipelineService.GetTrainingPipeline].
Required. The name of the TrainingPipeline resource. Format: `projects/{project}/locations/{location}/trainingPipelines/{training_pipeline}`
Lists TrainingPipelines in a Location.
Request message for [PipelineService.ListTrainingPipelines][google.cloud.aiplatform.v1.PipelineService.ListTrainingPipelines].
Required. The resource name of the Location to list the TrainingPipelines from. Format: `projects/{project}/locations/{location}`
The standard list filter. Supported fields: * `display_name` supports `=`, `!=` comparisons, and `:` wildcard. * `state` supports `=`, `!=` comparisons. * `training_task_definition` `=`, `!=` comparisons, and `:` wildcard. * `create_time` supports `=`, `!=`,`<`, `<=`,`>`, `>=` comparisons. `create_time` must be in RFC 3339 format. * `labels` supports general map functions that is: `labels.key=value` - key:value equality `labels.key:* - key existence Some examples of using the filter are: * `state="PIPELINE_STATE_SUCCEEDED" AND display_name:"my_pipeline_*"` * `state!="PIPELINE_STATE_FAILED" OR display_name="my_pipeline"` * `NOT display_name="my_pipeline"` * `create_time>"2021-05-18T00:00:00Z"` * `training_task_definition:"*automl_text_classification*"`
The standard list page size.
The standard list page token. Typically obtained via [ListTrainingPipelinesResponse.next_page_token][google.cloud.aiplatform.v1.ListTrainingPipelinesResponse.next_page_token] of the previous [PipelineService.ListTrainingPipelines][google.cloud.aiplatform.v1.PipelineService.ListTrainingPipelines] call.
Mask specifying which fields to read.
Response message for [PipelineService.ListTrainingPipelines][google.cloud.aiplatform.v1.PipelineService.ListTrainingPipelines]
List of TrainingPipelines in the requested page.
A token to retrieve the next page of results. Pass to [ListTrainingPipelinesRequest.page_token][google.cloud.aiplatform.v1.ListTrainingPipelinesRequest.page_token] to obtain that page.
Deletes a TrainingPipeline.
Request message for [PipelineService.DeleteTrainingPipeline][google.cloud.aiplatform.v1.PipelineService.DeleteTrainingPipeline].
Required. The name of the TrainingPipeline resource to be deleted. Format: `projects/{project}/locations/{location}/trainingPipelines/{training_pipeline}`
Cancels a TrainingPipeline. Starts asynchronous cancellation on the TrainingPipeline. The server makes a best effort to cancel the pipeline, but success is not guaranteed. Clients can use [PipelineService.GetTrainingPipeline][google.cloud.aiplatform.v1.PipelineService.GetTrainingPipeline] or other methods to check whether the cancellation succeeded or whether the pipeline completed despite cancellation. On successful cancellation, the TrainingPipeline is not deleted; instead it becomes a pipeline with a [TrainingPipeline.error][google.cloud.aiplatform.v1.TrainingPipeline.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [TrainingPipeline.state][google.cloud.aiplatform.v1.TrainingPipeline.state] is set to `CANCELLED`.
Request message for [PipelineService.CancelTrainingPipeline][google.cloud.aiplatform.v1.PipelineService.CancelTrainingPipeline].
Required. The name of the TrainingPipeline to cancel. Format: `projects/{project}/locations/{location}/trainingPipelines/{training_pipeline}`
Creates a PipelineJob. A PipelineJob will run immediately when created.
Gets a PipelineJob.
Request message for [PipelineService.GetPipelineJob][google.cloud.aiplatform.v1.PipelineService.GetPipelineJob].
Required. The name of the PipelineJob resource. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`
Lists PipelineJobs in a Location.
Request message for [PipelineService.ListPipelineJobs][google.cloud.aiplatform.v1.PipelineService.ListPipelineJobs].
Required. The resource name of the Location to list the PipelineJobs from. Format: `projects/{project}/locations/{location}`
Lists the PipelineJobs that match the filter expression. The following fields are supported: * `pipeline_name`: Supports `=` and `!=` comparisons. * `display_name`: Supports `=`, `!=` comparisons, and `:` wildcard. * `pipeline_job_user_id`: Supports `=`, `!=` comparisons, and `:` wildcard. for example, can check if pipeline's display_name contains *step* by doing display_name:\"*step*\" * `state`: Supports `=` and `!=` comparisons. * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `end_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality and key presence. * `template_uri`: Supports `=`, `!=` comparisons, and `:` wildcard. * `template_metadata.version`: Supports `=`, `!=` comparisons, and `:` wildcard. Filter expressions can be combined together using logical operators (`AND` & `OR`). For example: `pipeline_name="test" AND create_time>"2020-05-18T13:30:00Z"`. The syntax to define filter expression is based on https://google.aip.dev/160. Examples: * `create_time>"2021-05-18T00:00:00Z" OR update_time>"2020-05-18T00:00:00Z"` PipelineJobs created or updated after 2020-05-18 00:00:00 UTC. * `labels.env = "prod"` PipelineJobs with label "env" set to "prod".
The standard list page size.
The standard list page token. Typically obtained via [ListPipelineJobsResponse.next_page_token][google.cloud.aiplatform.v1.ListPipelineJobsResponse.next_page_token] of the previous [PipelineService.ListPipelineJobs][google.cloud.aiplatform.v1.PipelineService.ListPipelineJobs] call.
A comma-separated list of fields to order by. The default sort order is in ascending order. Use "desc" after a field name for descending. You can have multiple order_by fields provided e.g. "create_time desc, end_time", "end_time, start_time, update_time" For example, using "create_time desc, end_time" will order results by create time in descending order, and if there are multiple jobs having the same create time, order them by the end time in ascending order. if order_by is not specified, it will order by default order is create time in descending order. Supported fields: * `create_time` * `update_time` * `end_time` * `start_time`
Mask specifying which fields to read.
Response message for [PipelineService.ListPipelineJobs][google.cloud.aiplatform.v1.PipelineService.ListPipelineJobs]
List of PipelineJobs in the requested page.
A token to retrieve the next page of results. Pass to [ListPipelineJobsRequest.page_token][google.cloud.aiplatform.v1.ListPipelineJobsRequest.page_token] to obtain that page.
Deletes a PipelineJob.
Request message for [PipelineService.DeletePipelineJob][google.cloud.aiplatform.v1.PipelineService.DeletePipelineJob].
Required. The name of the PipelineJob resource to be deleted. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`
Batch deletes PipelineJobs The Operation is atomic. If it fails, none of the PipelineJobs are deleted. If it succeeds, all of the PipelineJobs are deleted.
Request message for [PipelineService.BatchDeletePipelineJobs][google.cloud.aiplatform.v1.PipelineService.BatchDeletePipelineJobs].
Required. The name of the PipelineJobs' parent resource. Format: `projects/{project}/locations/{location}`
Required. The names of the PipelineJobs to delete. A maximum of 32 PipelineJobs can be deleted in a batch. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipelineJob}`
Cancels a PipelineJob. Starts asynchronous cancellation on the PipelineJob. The server makes a best effort to cancel the pipeline, but success is not guaranteed. Clients can use [PipelineService.GetPipelineJob][google.cloud.aiplatform.v1.PipelineService.GetPipelineJob] or other methods to check whether the cancellation succeeded or whether the pipeline completed despite cancellation. On successful cancellation, the PipelineJob is not deleted; instead it becomes a pipeline with a [PipelineJob.error][google.cloud.aiplatform.v1.PipelineJob.error] value with a [google.rpc.Status.code][google.rpc.Status.code] of 1, corresponding to `Code.CANCELLED`, and [PipelineJob.state][google.cloud.aiplatform.v1.PipelineJob.state] is set to `CANCELLED`.
Request message for [PipelineService.CancelPipelineJob][google.cloud.aiplatform.v1.PipelineService.CancelPipelineJob].
Required. The name of the PipelineJob to cancel. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`
Batch cancel PipelineJobs. Firstly the server will check if all the jobs are in non-terminal states, and skip the jobs that are already terminated. If the operation failed, none of the pipeline jobs are cancelled. The server will poll the states of all the pipeline jobs periodically to check the cancellation status. This operation will return an LRO.
Request message for [PipelineService.BatchCancelPipelineJobs][google.cloud.aiplatform.v1.PipelineService.BatchCancelPipelineJobs].
Required. The name of the PipelineJobs' parent resource. Format: `projects/{project}/locations/{location}`
Required. The names of the PipelineJobs to cancel. A maximum of 32 PipelineJobs can be cancelled in a batch. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipelineJob}`
A service for online predictions and explanations.
Perform an online prediction.
Request message for [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' [Model's][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri].
The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' [Model's ][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [parameters_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri].
Response message for [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict].
The predictions that are the output of the predictions call. The schema of any single prediction may be specified via Endpoint's DeployedModels' [Model's ][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [prediction_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.prediction_schema_uri].
ID of the Endpoint's DeployedModel that served this prediction.
Output only. The resource name of the Model which is deployed as the DeployedModel that this prediction hits.
Output only. The version ID of the Model which is deployed as the DeployedModel that this prediction hits.
Output only. The [display name][google.cloud.aiplatform.v1.Model.display_name] of the Model which is deployed as the DeployedModel that this prediction hits.
Output only. Request-level metadata returned by the model. The metadata type will be dependent upon the model implementation.
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the [Endpoint][google.cloud.aiplatform.v1.Endpoint] that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's [DeployedModel][google.cloud.aiplatform.v1.DeployedModel] that served this prediction.
Request message for [PredictionService.RawPredict][google.cloud.aiplatform.v1.PredictionService.RawPredict].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
The prediction input. Supports HTTP headers and arbitrary data payload. A [DeployedModel][google.cloud.aiplatform.v1.DeployedModel] may have an upper limit on the number of instances it supports per request. When this limit it is exceeded for an AutoML model, the [RawPredict][google.cloud.aiplatform.v1.PredictionService.RawPredict] method returns an error. When this limit is exceeded for a custom-trained model, the behavior varies depending on the model. You can specify the schema for each instance in the [predict_schemata.instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] field when you create a [Model][google.cloud.aiplatform.v1.Model]. This schema applies when you deploy the `Model` as a `DeployedModel` to an [Endpoint][google.cloud.aiplatform.v1.Endpoint] and use the `RawPredict` method.
Perform a streaming online prediction with an arbitrary HTTP payload.
Request message for [PredictionService.StreamRawPredict][google.cloud.aiplatform.v1.PredictionService.StreamRawPredict].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
The prediction input. Supports HTTP headers and arbitrary data payload.
Perform an unary online prediction request to a gRPC model server for Vertex first-party products and frameworks.
Request message for [PredictionService.DirectPredict][google.cloud.aiplatform.v1.PredictionService.DirectPredict].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
The prediction input.
The parameters that govern the prediction.
Response message for [PredictionService.DirectPredict][google.cloud.aiplatform.v1.PredictionService.DirectPredict].
The prediction output.
The parameters that govern the prediction.
Perform an unary online prediction request to a gRPC model server for custom containers.
Request message for [PredictionService.DirectRawPredict][google.cloud.aiplatform.v1.PredictionService.DirectRawPredict].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Fully qualified name of the API method being invoked to perform predictions. Format: `/namespace.Service/Method/` Example: `/tensorflow.serving.PredictionService/Predict`
The prediction input.
Response message for [PredictionService.DirectRawPredict][google.cloud.aiplatform.v1.PredictionService.DirectRawPredict].
The prediction output.
Perform a streaming online prediction request to a gRPC model server for Vertex first-party products and frameworks.
Request message for [PredictionService.StreamDirectPredict][google.cloud.aiplatform.v1.PredictionService.StreamDirectPredict]. The first message must contain [endpoint][google.cloud.aiplatform.v1.StreamDirectPredictRequest.endpoint] field and optionally [input][]. The subsequent messages must contain [input][].
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Optional. The prediction input.
Optional. The parameters that govern the prediction.
Response message for [PredictionService.StreamDirectPredict][google.cloud.aiplatform.v1.PredictionService.StreamDirectPredict].
The prediction output.
The parameters that govern the prediction.
Perform a streaming online prediction request to a gRPC model server for custom containers.
Request message for [PredictionService.StreamDirectRawPredict][google.cloud.aiplatform.v1.PredictionService.StreamDirectRawPredict]. The first message must contain [endpoint][google.cloud.aiplatform.v1.StreamDirectRawPredictRequest.endpoint] and [method_name][google.cloud.aiplatform.v1.StreamDirectRawPredictRequest.method_name] fields and optionally [input][google.cloud.aiplatform.v1.StreamDirectRawPredictRequest.input]. The subsequent messages must contain [input][google.cloud.aiplatform.v1.StreamDirectRawPredictRequest.input]. [method_name][google.cloud.aiplatform.v1.StreamDirectRawPredictRequest.method_name] in the subsequent messages have no effect.
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Optional. Fully qualified name of the API method being invoked to perform predictions. Format: `/namespace.Service/Method/` Example: `/tensorflow.serving.PredictionService/Predict`
Optional. The prediction input.
Response message for [PredictionService.StreamDirectRawPredict][google.cloud.aiplatform.v1.PredictionService.StreamDirectRawPredict].
The prediction output.
Perform a streaming online prediction request for Vertex first-party products and frameworks.
Perform a server-side streaming online prediction request for Vertex LLM streaming.
Perform a streaming online prediction request through gRPC.
Request message for [PredictionService.StreamingRawPredict][google.cloud.aiplatform.v1.PredictionService.StreamingRawPredict]. The first message must contain [endpoint][google.cloud.aiplatform.v1.StreamingRawPredictRequest.endpoint] and [method_name][google.cloud.aiplatform.v1.StreamingRawPredictRequest.method_name] fields and optionally [input][google.cloud.aiplatform.v1.StreamingRawPredictRequest.input]. The subsequent messages must contain [input][google.cloud.aiplatform.v1.StreamingRawPredictRequest.input]. [method_name][google.cloud.aiplatform.v1.StreamingRawPredictRequest.method_name] in the subsequent messages have no effect.
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Fully qualified name of the API method being invoked to perform predictions. Format: `/namespace.Service/Method/` Example: `/tensorflow.serving.PredictionService/Predict`
The prediction input.
Response message for [PredictionService.StreamingRawPredict][google.cloud.aiplatform.v1.PredictionService.StreamingRawPredict].
The prediction output.
Perform an online explanation. If [deployed_model_id][google.cloud.aiplatform.v1.ExplainRequest.deployed_model_id] is specified, the corresponding DeployModel must have [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] populated. If [deployed_model_id][google.cloud.aiplatform.v1.ExplainRequest.deployed_model_id] is not specified, all DeployedModels must have [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] populated.
Request message for [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain].
Required. The name of the Endpoint requested to serve the explanation. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The instances that are the input to the explanation call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the explanation call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' [Model's][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri].
The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' [Model's ][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [parameters_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri].
If specified, overrides the [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] of the DeployedModel. Can be used for explaining prediction results with different configurations, such as: - Explaining top-5 predictions results as opposed to top-1; - Increasing path count or step count of the attribution methods to reduce approximate errors; - Using different baselines for explaining the prediction results.
If specified, this ExplainRequest will be served by the chosen DeployedModel, overriding [Endpoint.traffic_split][google.cloud.aiplatform.v1.Endpoint.traffic_split].
Response message for [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain].
The explanations of the Model's [PredictResponse.predictions][google.cloud.aiplatform.v1.PredictResponse.predictions]. It has the same number of elements as [instances][google.cloud.aiplatform.v1.ExplainRequest.instances] to be explained.
ID of the Endpoint's DeployedModel that served this explanation.
The predictions that are the output of the predictions call. Same as [PredictResponse.predictions][google.cloud.aiplatform.v1.PredictResponse.predictions].
Generate content with multimodal inputs.
Generate content with multimodal inputs with streaming support.
A service for executing queries on Reasoning Engine.
Queries using a reasoning engine.
Request message for [ReasoningEngineExecutionService.Query][].
Required. The name of the ReasoningEngine resource to use. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}`
Optional. Input content provided by users in JSON object format. Examples include text query, function calling parameters, media bytes, etc.
Optional. Class method to be used for the query. It is optional and defaults to "query" if unspecified.
Response message for [ReasoningEngineExecutionService.Query][]
Response provided by users in JSON object format.
Streams queries using a reasoning engine.
Request message for [ReasoningEngineExecutionService.StreamQuery][].
Required. The name of the ReasoningEngine resource to use. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}`
Optional. Input content provided by users in JSON object format. Examples include text query, function calling parameters, media bytes, etc.
Optional. Class method to be used for the stream query. It is optional and defaults to "stream_query" if unspecified.
A service for managing Vertex AI's Reasoning Engines.
Creates a reasoning engine.
Request message for [ReasoningEngineService.CreateReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.CreateReasoningEngine].
Required. The resource name of the Location to create the ReasoningEngine in. Format: `projects/{project}/locations/{location}`
Required. The ReasoningEngine to create.
Gets a reasoning engine.
Request message for [ReasoningEngineService.GetReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.GetReasoningEngine].
Required. The name of the ReasoningEngine resource. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}`
Lists reasoning engines in a location.
Request message for [ReasoningEngineService.ListReasoningEngines][google.cloud.aiplatform.v1.ReasoningEngineService.ListReasoningEngines].
Required. The resource name of the Location to list the ReasoningEngines from. Format: `projects/{project}/locations/{location}`
Optional. The standard list filter. More detail in [AIP-160](https://google.aip.dev/160).
Optional. The standard list page size.
Optional. The standard list page token.
Response message for [ReasoningEngineService.ListReasoningEngines][google.cloud.aiplatform.v1.ReasoningEngineService.ListReasoningEngines]
List of ReasoningEngines in the requested page.
A token to retrieve the next page of results. Pass to [ListReasoningEnginesRequest.page_token][google.cloud.aiplatform.v1.ListReasoningEnginesRequest.page_token] to obtain that page.
Updates a reasoning engine.
Request message for [ReasoningEngineService.UpdateReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.UpdateReasoningEngine].
Required. The ReasoningEngine which replaces the resource on the server.
Optional. Mask specifying which fields to update.
Deletes a reasoning engine.
Request message for [ReasoningEngineService.DeleteReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.DeleteReasoningEngine].
Required. The name of the ReasoningEngine resource to be deleted. Format: `projects/{project}/locations/{location}/reasoningEngines/{reasoning_engine}`
Optional. If set to true, child resources of this reasoning engine will also be deleted. Otherwise, the request will fail with FAILED_PRECONDITION error when the reasoning engine has undeleted child resources.
A service for creating and managing Vertex AI's Schedule resources to periodically launch shceudled runs to make API calls.
Creates a Schedule.
Request message for [ScheduleService.CreateSchedule][google.cloud.aiplatform.v1.ScheduleService.CreateSchedule].
Required. The resource name of the Location to create the Schedule in. Format: `projects/{project}/locations/{location}`
Required. The Schedule to create.
Deletes a Schedule.
Request message for [ScheduleService.DeleteSchedule][google.cloud.aiplatform.v1.ScheduleService.DeleteSchedule].
Required. The name of the Schedule resource to be deleted. Format: `projects/{project}/locations/{location}/schedules/{schedule}`
Gets a Schedule.
Request message for [ScheduleService.GetSchedule][google.cloud.aiplatform.v1.ScheduleService.GetSchedule].
Required. The name of the Schedule resource. Format: `projects/{project}/locations/{location}/schedules/{schedule}`
Lists Schedules in a Location.
Request message for [ScheduleService.ListSchedules][google.cloud.aiplatform.v1.ScheduleService.ListSchedules].
Required. The resource name of the Location to list the Schedules from. Format: `projects/{project}/locations/{location}`
Lists the Schedules that match the filter expression. The following fields are supported: * `display_name`: Supports `=`, `!=` comparisons, and `:` wildcard. * `state`: Supports `=` and `!=` comparisons. * `request`: Supports existence of the <request_type> check. (e.g. `create_pipeline_job_request:*` --> Schedule has create_pipeline_job_request). * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `start_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `end_time`: Supports `=`, `!=`, `<`, `>`, `<=`, `>=` comparisons and `:*` existence check. Values must be in RFC 3339 format. * `next_run_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. Filter expressions can be combined together using logical operators (`NOT`, `AND` & `OR`). The syntax to define filter expression is based on https://google.aip.dev/160. Examples: * `state="ACTIVE" AND display_name:"my_schedule_*"` * `NOT display_name="my_schedule"` * `create_time>"2021-05-18T00:00:00Z"` * `end_time>"2021-05-18T00:00:00Z" OR NOT end_time:*` * `create_pipeline_job_request:*`
The standard list page size. Default to 100 if not specified.
The standard list page token. Typically obtained via [ListSchedulesResponse.next_page_token][google.cloud.aiplatform.v1.ListSchedulesResponse.next_page_token] of the previous [ScheduleService.ListSchedules][google.cloud.aiplatform.v1.ScheduleService.ListSchedules] call.
A comma-separated list of fields to order by. The default sort order is in ascending order. Use "desc" after a field name for descending. You can have multiple order_by fields provided. For example, using "create_time desc, end_time" will order results by create time in descending order, and if there are multiple schedules having the same create time, order them by the end time in ascending order. If order_by is not specified, it will order by default with create_time in descending order. Supported fields: * `create_time` * `start_time` * `end_time` * `next_run_time`
Response message for [ScheduleService.ListSchedules][google.cloud.aiplatform.v1.ScheduleService.ListSchedules]
List of Schedules in the requested page.
A token to retrieve the next page of results. Pass to [ListSchedulesRequest.page_token][google.cloud.aiplatform.v1.ListSchedulesRequest.page_token] to obtain that page.
Pauses a Schedule. Will mark [Schedule.state][google.cloud.aiplatform.v1.Schedule.state] to 'PAUSED'. If the schedule is paused, no new runs will be created. Already created runs will NOT be paused or canceled.
Request message for [ScheduleService.PauseSchedule][google.cloud.aiplatform.v1.ScheduleService.PauseSchedule].
Required. The name of the Schedule resource to be paused. Format: `projects/{project}/locations/{location}/schedules/{schedule}`
Resumes a paused Schedule to start scheduling new runs. Will mark [Schedule.state][google.cloud.aiplatform.v1.Schedule.state] to 'ACTIVE'. Only paused Schedule can be resumed. When the Schedule is resumed, new runs will be scheduled starting from the next execution time after the current time based on the time_specification in the Schedule. If [Schedule.catch_up][google.cloud.aiplatform.v1.Schedule.catch_up] is set up true, all missed runs will be scheduled for backfill first.
Request message for [ScheduleService.ResumeSchedule][google.cloud.aiplatform.v1.ScheduleService.ResumeSchedule].
Required. The name of the Schedule resource to be resumed. Format: `projects/{project}/locations/{location}/schedules/{schedule}`
Optional. Whether to backfill missed runs when the schedule is resumed from PAUSED state. If set to true, all missed runs will be scheduled. New runs will be scheduled after the backfill is complete. This will also update [Schedule.catch_up][google.cloud.aiplatform.v1.Schedule.catch_up] field. Default to false.
Updates an active or paused Schedule. When the Schedule is updated, new runs will be scheduled starting from the updated next execution time after the update time based on the time_specification in the updated Schedule. All unstarted runs before the update time will be skipped while already created runs will NOT be paused or canceled.
Request message for [ScheduleService.UpdateSchedule][google.cloud.aiplatform.v1.ScheduleService.UpdateSchedule].
Required. The Schedule which replaces the resource on the server. The following restrictions will be applied: * The scheduled request type cannot be changed. * The non-empty fields cannot be unset. * The output_only fields will be ignored if specified.
Required. The update mask applies to the resource. See [google.protobuf.FieldMask][google.protobuf.FieldMask].
A service for creating and managing Customer SpecialistPools. When customers start Data Labeling jobs, they can reuse/create Specialist Pools to bring their own Specialists to label the data. Customers can add/remove Managers for the Specialist Pool on Cloud console, then Managers will get email notifications to manage Specialists and tasks on CrowdCompute console.
Creates a SpecialistPool.
Request message for [SpecialistPoolService.CreateSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.CreateSpecialistPool].
Required. The parent Project name for the new SpecialistPool. The form is `projects/{project}/locations/{location}`.
Required. The SpecialistPool to create.
Gets a SpecialistPool.
Request message for [SpecialistPoolService.GetSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.GetSpecialistPool].
Required. The name of the SpecialistPool resource. The form is `projects/{project}/locations/{location}/specialistPools/{specialist_pool}`.
Lists SpecialistPools in a Location.
Request message for [SpecialistPoolService.ListSpecialistPools][google.cloud.aiplatform.v1.SpecialistPoolService.ListSpecialistPools].
Required. The name of the SpecialistPool's parent resource. Format: `projects/{project}/locations/{location}`
The standard list page size.
The standard list page token. Typically obtained by [ListSpecialistPoolsResponse.next_page_token][google.cloud.aiplatform.v1.ListSpecialistPoolsResponse.next_page_token] of the previous [SpecialistPoolService.ListSpecialistPools][google.cloud.aiplatform.v1.SpecialistPoolService.ListSpecialistPools] call. Return first page if empty.
Mask specifying which fields to read. FieldMask represents a set of
Response message for [SpecialistPoolService.ListSpecialistPools][google.cloud.aiplatform.v1.SpecialistPoolService.ListSpecialistPools].
A list of SpecialistPools that matches the specified filter in the request.
The standard List next-page token.
Deletes a SpecialistPool as well as all Specialists in the pool.
Request message for [SpecialistPoolService.DeleteSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.DeleteSpecialistPool].
Required. The resource name of the SpecialistPool to delete. Format: `projects/{project}/locations/{location}/specialistPools/{specialist_pool}`
If set to true, any specialist managers in this SpecialistPool will also be deleted. (Otherwise, the request will only work if the SpecialistPool has no specialist managers.)
Updates a SpecialistPool.
Request message for [SpecialistPoolService.UpdateSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.UpdateSpecialistPool].
Required. The SpecialistPool which replaces the resource on the server.
Required. The update mask applies to the resource.
TensorboardService
Creates a Tensorboard.
Request message for [TensorboardService.CreateTensorboard][google.cloud.aiplatform.v1.TensorboardService.CreateTensorboard].
Required. The resource name of the Location to create the Tensorboard in. Format: `projects/{project}/locations/{location}`
Required. The Tensorboard to create.
Gets a Tensorboard.
Request message for [TensorboardService.GetTensorboard][google.cloud.aiplatform.v1.TensorboardService.GetTensorboard].
Required. The name of the Tensorboard resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Updates a Tensorboard.
Request message for [TensorboardService.UpdateTensorboard][google.cloud.aiplatform.v1.TensorboardService.UpdateTensorboard].
Required. Field mask is used to specify the fields to be overwritten in the Tensorboard resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified.
Required. The Tensorboard's `name` field is used to identify the Tensorboard to be updated. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Lists Tensorboards in a Location.
Request message for [TensorboardService.ListTensorboards][google.cloud.aiplatform.v1.TensorboardService.ListTensorboards].
Required. The resource name of the Location to list Tensorboards. Format: `projects/{project}/locations/{location}`
Lists the Tensorboards that match the filter expression.
The maximum number of Tensorboards to return. The service may return fewer than this value. If unspecified, at most 100 Tensorboards are returned. The maximum value is 100; values above 100 are coerced to 100.
A page token, received from a previous [TensorboardService.ListTensorboards][google.cloud.aiplatform.v1.TensorboardService.ListTensorboards] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [TensorboardService.ListTensorboards][google.cloud.aiplatform.v1.TensorboardService.ListTensorboards] must match the call that provided the page token.
Field to use to sort the list.
Mask specifying which fields to read.
Response message for [TensorboardService.ListTensorboards][google.cloud.aiplatform.v1.TensorboardService.ListTensorboards].
The Tensorboards mathching the request.
A token, which can be sent as [ListTensorboardsRequest.page_token][google.cloud.aiplatform.v1.ListTensorboardsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Deletes a Tensorboard.
Request message for [TensorboardService.DeleteTensorboard][google.cloud.aiplatform.v1.TensorboardService.DeleteTensorboard].
Required. The name of the Tensorboard to be deleted. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Returns a list of monthly active users for a given TensorBoard instance.
Request message for [TensorboardService.ReadTensorboardUsage][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardUsage].
Required. The name of the Tensorboard resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Response message for [TensorboardService.ReadTensorboardUsage][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardUsage].
Maps year-month (YYYYMM) string to per month usage data.
Returns the storage size for a given TensorBoard instance.
Request message for [TensorboardService.ReadTensorboardSize][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardSize].
Required. The name of the Tensorboard resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Response message for [TensorboardService.ReadTensorboardSize][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardSize].
Payload storage size for the TensorBoard
Creates a TensorboardExperiment.
Request message for [TensorboardService.CreateTensorboardExperiment][google.cloud.aiplatform.v1.TensorboardService.CreateTensorboardExperiment].
Required. The resource name of the Tensorboard to create the TensorboardExperiment in. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
The TensorboardExperiment to create.
Required. The ID to use for the Tensorboard experiment, which becomes the final component of the Tensorboard experiment's resource name. This value should be 1-128 characters, and valid characters are `/[a-z][0-9]-/`.
Gets a TensorboardExperiment.
Request message for [TensorboardService.GetTensorboardExperiment][google.cloud.aiplatform.v1.TensorboardService.GetTensorboardExperiment].
Required. The name of the TensorboardExperiment resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Updates a TensorboardExperiment.
Request message for [TensorboardService.UpdateTensorboardExperiment][google.cloud.aiplatform.v1.TensorboardService.UpdateTensorboardExperiment].
Required. Field mask is used to specify the fields to be overwritten in the TensorboardExperiment resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified.
Required. The TensorboardExperiment's `name` field is used to identify the TensorboardExperiment to be updated. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Lists TensorboardExperiments in a Location.
Request message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments].
Required. The resource name of the Tensorboard to list TensorboardExperiments. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Lists the TensorboardExperiments that match the filter expression.
The maximum number of TensorboardExperiments to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardExperiments are returned. The maximum value is 1000; values above 1000 are coerced to 1000.
A page token, received from a previous [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments] must match the call that provided the page token.
Field to use to sort the list.
Mask specifying which fields to read.
Response message for [TensorboardService.ListTensorboardExperiments][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardExperiments].
The TensorboardExperiments mathching the request.
A token, which can be sent as [ListTensorboardExperimentsRequest.page_token][google.cloud.aiplatform.v1.ListTensorboardExperimentsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Deletes a TensorboardExperiment.
Request message for [TensorboardService.DeleteTensorboardExperiment][google.cloud.aiplatform.v1.TensorboardService.DeleteTensorboardExperiment].
Required. The name of the TensorboardExperiment to be deleted. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Creates a TensorboardRun.
Batch create TensorboardRuns.
Request message for [TensorboardService.BatchCreateTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.BatchCreateTensorboardRuns].
Required. The resource name of the TensorboardExperiment to create the TensorboardRuns in. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}` The parent field in the CreateTensorboardRunRequest messages must match this field.
Required. The request message specifying the TensorboardRuns to create. A maximum of 1000 TensorboardRuns can be created in a batch.
Response message for [TensorboardService.BatchCreateTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.BatchCreateTensorboardRuns].
The created TensorboardRuns.
Gets a TensorboardRun.
Request message for [TensorboardService.GetTensorboardRun][google.cloud.aiplatform.v1.TensorboardService.GetTensorboardRun].
Required. The name of the TensorboardRun resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Updates a TensorboardRun.
Request message for [TensorboardService.UpdateTensorboardRun][google.cloud.aiplatform.v1.TensorboardService.UpdateTensorboardRun].
Required. Field mask is used to specify the fields to be overwritten in the TensorboardRun resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified.
Required. The TensorboardRun's `name` field is used to identify the TensorboardRun to be updated. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Lists TensorboardRuns in a Location.
Request message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns].
Required. The resource name of the TensorboardExperiment to list TensorboardRuns. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Lists the TensorboardRuns that match the filter expression.
The maximum number of TensorboardRuns to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardRuns are returned. The maximum value is 1000; values above 1000 are coerced to 1000.
A page token, received from a previous [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns] must match the call that provided the page token.
Field to use to sort the list.
Mask specifying which fields to read.
Response message for [TensorboardService.ListTensorboardRuns][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardRuns].
The TensorboardRuns mathching the request.
A token, which can be sent as [ListTensorboardRunsRequest.page_token][google.cloud.aiplatform.v1.ListTensorboardRunsRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Deletes a TensorboardRun.
Request message for [TensorboardService.DeleteTensorboardRun][google.cloud.aiplatform.v1.TensorboardService.DeleteTensorboardRun].
Required. The name of the TensorboardRun to be deleted. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Batch create TensorboardTimeSeries that belong to a TensorboardExperiment.
Request message for [TensorboardService.BatchCreateTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.BatchCreateTensorboardTimeSeries].
Required. The resource name of the TensorboardExperiment to create the TensorboardTimeSeries in. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}` The TensorboardRuns referenced by the parent fields in the CreateTensorboardTimeSeriesRequest messages must be sub resources of this TensorboardExperiment.
Required. The request message specifying the TensorboardTimeSeries to create. A maximum of 1000 TensorboardTimeSeries can be created in a batch.
Response message for [TensorboardService.BatchCreateTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.BatchCreateTensorboardTimeSeries].
The created TensorboardTimeSeries.
Creates a TensorboardTimeSeries.
Gets a TensorboardTimeSeries.
Request message for [TensorboardService.GetTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.GetTensorboardTimeSeries].
Required. The name of the TensorboardTimeSeries resource. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
Updates a TensorboardTimeSeries.
Request message for [TensorboardService.UpdateTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.UpdateTensorboardTimeSeries].
Required. Field mask is used to specify the fields to be overwritten in the TensorboardTimeSeries resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field is overwritten if it's in the mask. If the user does not provide a mask then all fields are overwritten if new values are specified.
Required. The TensorboardTimeSeries' `name` field is used to identify the TensorboardTimeSeries to be updated. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
Lists TensorboardTimeSeries in a Location.
Request message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries].
Required. The resource name of the TensorboardRun to list TensorboardTimeSeries. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Lists the TensorboardTimeSeries that match the filter expression.
The maximum number of TensorboardTimeSeries to return. The service may return fewer than this value. If unspecified, at most 50 TensorboardTimeSeries are returned. The maximum value is 1000; values above 1000 are coerced to 1000.
A page token, received from a previous [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries] must match the call that provided the page token.
Field to use to sort the list.
Mask specifying which fields to read.
Response message for [TensorboardService.ListTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.ListTensorboardTimeSeries].
The TensorboardTimeSeries mathching the request.
A token, which can be sent as [ListTensorboardTimeSeriesRequest.page_token][google.cloud.aiplatform.v1.ListTensorboardTimeSeriesRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Deletes a TensorboardTimeSeries.
Request message for [TensorboardService.DeleteTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.DeleteTensorboardTimeSeries].
Required. The name of the TensorboardTimeSeries to be deleted. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
Reads multiple TensorboardTimeSeries' data. The data point number limit is 1000 for scalars, 100 for tensors and blob references. If the number of data points stored is less than the limit, all data is returned. Otherwise, the number limit of data points is randomly selected from this time series and returned.
Request message for [TensorboardService.BatchReadTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.BatchReadTensorboardTimeSeriesData].
Required. The resource name of the Tensorboard containing TensorboardTimeSeries to read data from. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`. The TensorboardTimeSeries referenced by [time_series][google.cloud.aiplatform.v1.BatchReadTensorboardTimeSeriesDataRequest.time_series] must be sub resources of this Tensorboard.
Required. The resource names of the TensorboardTimeSeries to read data from. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
Response message for [TensorboardService.BatchReadTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.BatchReadTensorboardTimeSeriesData].
The returned time series data.
Reads a TensorboardTimeSeries' data. By default, if the number of data points stored is less than 1000, all data is returned. Otherwise, 1000 data points is randomly selected from this time series and returned. This value can be changed by changing max_data_points, which can't be greater than 10k.
Request message for [TensorboardService.ReadTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardTimeSeriesData].
Required. The resource name of the TensorboardTimeSeries to read data from. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
The maximum number of TensorboardTimeSeries' data to return. This value should be a positive integer. This value can be set to -1 to return all data.
Reads the TensorboardTimeSeries' data that match the filter expression.
Response message for [TensorboardService.ReadTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardTimeSeriesData].
The returned time series data.
Gets bytes of TensorboardBlobs. This is to allow reading blob data stored in consumer project's Cloud Storage bucket without users having to obtain Cloud Storage access permission.
Request message for [TensorboardService.ReadTensorboardBlobData][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardBlobData].
Required. The resource name of the TensorboardTimeSeries to list Blobs. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
IDs of the blobs to read.
Response message for [TensorboardService.ReadTensorboardBlobData][google.cloud.aiplatform.v1.TensorboardService.ReadTensorboardBlobData].
Blob messages containing blob bytes.
Write time series data points of multiple TensorboardTimeSeries in multiple TensorboardRun's. If any data fail to be ingested, an error is returned.
Request message for [TensorboardService.WriteTensorboardExperimentData][google.cloud.aiplatform.v1.TensorboardService.WriteTensorboardExperimentData].
Required. The resource name of the TensorboardExperiment to write data to. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Required. Requests containing per-run TensorboardTimeSeries data to write.
Response message for [TensorboardService.WriteTensorboardExperimentData][google.cloud.aiplatform.v1.TensorboardService.WriteTensorboardExperimentData].
(message has no fields)
Write time series data points into multiple TensorboardTimeSeries under a TensorboardRun. If any data fail to be ingested, an error is returned.
Response message for [TensorboardService.WriteTensorboardRunData][google.cloud.aiplatform.v1.TensorboardService.WriteTensorboardRunData].
(message has no fields)
Exports a TensorboardTimeSeries' data. Data is returned in paginated responses.
Request message for [TensorboardService.ExportTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ExportTensorboardTimeSeriesData].
Required. The resource name of the TensorboardTimeSeries to export data from. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}/timeSeries/{time_series}`
Exports the TensorboardTimeSeries' data that match the filter expression.
The maximum number of data points to return per page. The default page_size is 1000. Values must be between 1 and 10000. Values above 10000 are coerced to 10000.
A page token, received from a previous [ExportTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ExportTensorboardTimeSeriesData] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [ExportTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ExportTensorboardTimeSeriesData] must match the call that provided the page token.
Field to use to sort the TensorboardTimeSeries' data. By default, TensorboardTimeSeries' data is returned in a pseudo random order.
Response message for [TensorboardService.ExportTensorboardTimeSeriesData][google.cloud.aiplatform.v1.TensorboardService.ExportTensorboardTimeSeriesData].
The returned time series data points.
A token, which can be sent as [page_token][google.cloud.aiplatform.v1.ExportTensorboardTimeSeriesDataRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
A service for managing user data for RAG.
Creates a RagCorpus.
Request message for [VertexRagDataService.CreateRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.CreateRagCorpus].
Required. The resource name of the Location to create the RagCorpus in. Format: `projects/{project}/locations/{location}`
Required. The RagCorpus to create.
Updates a RagCorpus.
Request message for [VertexRagDataService.UpdateRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.UpdateRagCorpus].
Required. The RagCorpus which replaces the resource on the server.
Gets a RagCorpus.
Request message for [VertexRagDataService.GetRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.GetRagCorpus]
Required. The name of the RagCorpus resource. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Lists RagCorpora in a Location.
Request message for [VertexRagDataService.ListRagCorpora][google.cloud.aiplatform.v1.VertexRagDataService.ListRagCorpora].
Required. The resource name of the Location from which to list the RagCorpora. Format: `projects/{project}/locations/{location}`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListRagCorporaResponse.next_page_token][google.cloud.aiplatform.v1.ListRagCorporaResponse.next_page_token] of the previous [VertexRagDataService.ListRagCorpora][google.cloud.aiplatform.v1.VertexRagDataService.ListRagCorpora] call.
Response message for [VertexRagDataService.ListRagCorpora][google.cloud.aiplatform.v1.VertexRagDataService.ListRagCorpora].
List of RagCorpora in the requested page.
A token to retrieve the next page of results. Pass to [ListRagCorporaRequest.page_token][google.cloud.aiplatform.v1.ListRagCorporaRequest.page_token] to obtain that page.
Deletes a RagCorpus.
Request message for [VertexRagDataService.DeleteRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.DeleteRagCorpus].
Required. The name of the RagCorpus resource to be deleted. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Optional. If set to true, any RagFiles in this RagCorpus will also be deleted. Otherwise, the request will only work if the RagCorpus has no RagFiles.
Upload a file into a RagCorpus.
Request message for [VertexRagDataService.UploadRagFile][google.cloud.aiplatform.v1.VertexRagDataService.UploadRagFile].
Required. The name of the RagCorpus resource into which to upload the file. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Required. The RagFile to upload.
Required. The config for the RagFiles to be uploaded into the RagCorpus. [VertexRagDataService.UploadRagFile][google.cloud.aiplatform.v1.VertexRagDataService.UploadRagFile].
Response message for [VertexRagDataService.UploadRagFile][google.cloud.aiplatform.v1.VertexRagDataService.UploadRagFile].
The result of the upload.
The RagFile that had been uploaded into the RagCorpus.
The error that occurred while processing the RagFile.
Import files from Google Cloud Storage or Google Drive into a RagCorpus.
Request message for [VertexRagDataService.ImportRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ImportRagFiles].
Required. The name of the RagCorpus resource into which to import files. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Required. The config for the RagFiles to be synced and imported into the RagCorpus. [VertexRagDataService.ImportRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ImportRagFiles].
Gets a RagFile.
Request message for [VertexRagDataService.GetRagFile][google.cloud.aiplatform.v1.VertexRagDataService.GetRagFile]
Required. The name of the RagFile resource. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}/ragFiles/{rag_file}`
Lists RagFiles in a RagCorpus.
Request message for [VertexRagDataService.ListRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ListRagFiles].
Required. The resource name of the RagCorpus from which to list the RagFiles. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Optional. The standard list page size.
Optional. The standard list page token. Typically obtained via [ListRagFilesResponse.next_page_token][google.cloud.aiplatform.v1.ListRagFilesResponse.next_page_token] of the previous [VertexRagDataService.ListRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ListRagFiles] call.
Response message for [VertexRagDataService.ListRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ListRagFiles].
List of RagFiles in the requested page.
A token to retrieve the next page of results. Pass to [ListRagFilesRequest.page_token][google.cloud.aiplatform.v1.ListRagFilesRequest.page_token] to obtain that page.
Deletes a RagFile.
Request message for [VertexRagDataService.DeleteRagFile][google.cloud.aiplatform.v1.VertexRagDataService.DeleteRagFile].
Required. The name of the RagFile resource to be deleted. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}/ragFiles/{rag_file}`
A service for retrieving relevant contexts.
Retrieves relevant contexts for a query.
Request message for [VertexRagService.RetrieveContexts][google.cloud.aiplatform.v1.VertexRagService.RetrieveContexts].
Data Source to retrieve contexts.
The data source for Vertex RagStore.
Required. The resource name of the Location from which to retrieve RagContexts. The users must have permission to make a call in the project. Format: `projects/{project}/locations/{location}`.
Required. Single RAG retrieve query.
Response message for [VertexRagService.RetrieveContexts][google.cloud.aiplatform.v1.VertexRagService.RetrieveContexts].
The contexts of the query.
Given an input prompt, it returns augmented prompt from vertex rag store to guide LLM towards generating grounded responses.
Request message for AugmentPrompt.
The data source for retrieving contexts.
Optional. Retrieves contexts from the Vertex RagStore.
Required. The resource name of the Location from which to augment prompt. The users must have permission to make a call in the project. Format: `projects/{project}/locations/{location}`.
Optional. Input content to augment, only text format is supported for now.
Optional. Metadata of the backend deployed model.
Response message for AugmentPrompt.
Augmented prompt, only text format is supported for now.
Retrieved facts from RAG data sources.
Given an input text, it returns a score that evaluates the factuality of the text. It also extracts and returns claims from the text and provides supporting facts.
Request message for CorroborateContent.
Required. The resource name of the Location from which to corroborate text. The users must have permission to make a call in the project. Format: `projects/{project}/locations/{location}`.
Optional. Input content to corroborate, only text format is supported for now.
Optional. Facts used to generate the text can also be used to corroborate the text.
Optional. Parameters that can be set to override default settings per request.
Response message for CorroborateContent.
Confidence score of corroborating content. Value is [0,1] with 1 is the most confidence.
Claims that are extracted from the input content and facts that support the claims.
Vertex AI Vizier API. Vertex AI Vizier is a service to solve blackbox optimization problems, such as tuning machine learning hyperparameters and searching over deep learning architectures.
Creates a Study. A resource name will be generated after creation of the Study.
Request message for [VizierService.CreateStudy][google.cloud.aiplatform.v1.VizierService.CreateStudy].
Required. The resource name of the Location to create the CustomJob in. Format: `projects/{project}/locations/{location}`
Required. The Study configuration used to create the Study.
Gets a Study by name.
Request message for [VizierService.GetStudy][google.cloud.aiplatform.v1.VizierService.GetStudy].
Required. The name of the Study resource. Format: `projects/{project}/locations/{location}/studies/{study}`
Lists all the studies in a region for an associated project.
Request message for [VizierService.ListStudies][google.cloud.aiplatform.v1.VizierService.ListStudies].
Required. The resource name of the Location to list the Study from. Format: `projects/{project}/locations/{location}`
Optional. A page token to request the next page of results. If unspecified, there are no subsequent pages.
Optional. The maximum number of studies to return per "page" of results. If unspecified, service will pick an appropriate default.
Response message for [VizierService.ListStudies][google.cloud.aiplatform.v1.VizierService.ListStudies].
The studies associated with the project.
Passes this token as the `page_token` field of the request for a subsequent call. If this field is omitted, there are no subsequent pages.
Deletes a Study.
Request message for [VizierService.DeleteStudy][google.cloud.aiplatform.v1.VizierService.DeleteStudy].
Required. The name of the Study resource to be deleted. Format: `projects/{project}/locations/{location}/studies/{study}`
Looks a study up using the user-defined display_name field instead of the fully qualified resource name.
Request message for [VizierService.LookupStudy][google.cloud.aiplatform.v1.VizierService.LookupStudy].
Required. The resource name of the Location to get the Study from. Format: `projects/{project}/locations/{location}`
Required. The user-defined display name of the Study
Adds one or more Trials to a Study, with parameter values suggested by Vertex AI Vizier. Returns a long-running operation associated with the generation of Trial suggestions. When this long-running operation succeeds, it will contain a [SuggestTrialsResponse][google.cloud.aiplatform.v1.SuggestTrialsResponse].
Request message for [VizierService.SuggestTrials][google.cloud.aiplatform.v1.VizierService.SuggestTrials].
Required. The project and location that the Study belongs to. Format: `projects/{project}/locations/{location}/studies/{study}`
Required. The number of suggestions requested. It must be positive.
Required. The identifier of the client that is requesting the suggestion. If multiple SuggestTrialsRequests have the same `client_id`, the service will return the identical suggested Trial if the Trial is pending, and provide a new Trial if the last suggested Trial was completed.
Optional. This allows you to specify the "context" for a Trial; a context is a slice (a subspace) of the search space. Typical uses for contexts: 1) You are using Vizier to tune a server for best performance, but there's a strong weekly cycle. The context specifies the day-of-week. This allows Tuesday to generalize from Wednesday without assuming that everything is identical. 2) Imagine you're optimizing some medical treatment for people. As they walk in the door, you know certain facts about them (e.g. sex, weight, height, blood-pressure). Put that information in the context, and Vizier will adapt its suggestions to the patient. 3) You want to do a fair A/B test efficiently. Specify the "A" and "B" conditions as contexts, and Vizier will generalize between "A" and "B" conditions. If they are similar, this will allow Vizier to converge to the optimum faster than if "A" and "B" were separate Studies. NOTE: You can also enter contexts as REQUESTED Trials, e.g. via the CreateTrial() RPC; that's the asynchronous option where you don't need a close association between contexts and suggestions. NOTE: All the Parameters you set in a context MUST be defined in the Study. NOTE: You must supply 0 or $suggestion_count contexts. If you don't supply any contexts, Vizier will make suggestions from the full search space specified in the StudySpec; if you supply a full set of context, each suggestion will match the corresponding context. NOTE: A Context with no features set matches anything, and allows suggestions from the full search space. NOTE: Contexts MUST lie within the search space specified in the StudySpec. It's an error if they don't. NOTE: Contexts preferentially match ACTIVE then REQUESTED trials before new suggestions are generated. NOTE: Generation of suggestions involves a match between a Context and (optionally) a REQUESTED trial; if that match is not fully specified, a suggestion will be geneated in the merged subspace.
Adds a user provided Trial to a Study.
Request message for [VizierService.CreateTrial][google.cloud.aiplatform.v1.VizierService.CreateTrial].
Required. The resource name of the Study to create the Trial in. Format: `projects/{project}/locations/{location}/studies/{study}`
Required. The Trial to create.
Gets a Trial.
Request message for [VizierService.GetTrial][google.cloud.aiplatform.v1.VizierService.GetTrial].
Required. The name of the Trial resource. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Lists the Trials associated with a Study.
Request message for [VizierService.ListTrials][google.cloud.aiplatform.v1.VizierService.ListTrials].
Required. The resource name of the Study to list the Trial from. Format: `projects/{project}/locations/{location}/studies/{study}`
Optional. A page token to request the next page of results. If unspecified, there are no subsequent pages.
Optional. The number of Trials to retrieve per "page" of results. If unspecified, the service will pick an appropriate default.
Response message for [VizierService.ListTrials][google.cloud.aiplatform.v1.VizierService.ListTrials].
The Trials associated with the Study.
Pass this token as the `page_token` field of the request for a subsequent call. If this field is omitted, there are no subsequent pages.
Adds a measurement of the objective metrics to a Trial. This measurement is assumed to have been taken before the Trial is complete.
Request message for [VizierService.AddTrialMeasurement][google.cloud.aiplatform.v1.VizierService.AddTrialMeasurement].
Required. The name of the trial to add measurement. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Required. The measurement to be added to a Trial.
Marks a Trial as complete.
Request message for [VizierService.CompleteTrial][google.cloud.aiplatform.v1.VizierService.CompleteTrial].
Required. The Trial's name. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Optional. If provided, it will be used as the completed Trial's final_measurement; Otherwise, the service will auto-select a previously reported measurement as the final-measurement
Optional. True if the Trial cannot be run with the given Parameter, and final_measurement will be ignored.
Optional. A human readable reason why the trial was infeasible. This should only be provided if `trial_infeasible` is true.
Deletes a Trial.
Request message for [VizierService.DeleteTrial][google.cloud.aiplatform.v1.VizierService.DeleteTrial].
Required. The Trial's name. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Checks whether a Trial should stop or not. Returns a long-running operation. When the operation is successful, it will contain a [CheckTrialEarlyStoppingStateResponse][google.cloud.aiplatform.v1.CheckTrialEarlyStoppingStateResponse].
Request message for [VizierService.CheckTrialEarlyStoppingState][google.cloud.aiplatform.v1.VizierService.CheckTrialEarlyStoppingState].
Required. The Trial's name. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Stops a Trial.
Request message for [VizierService.StopTrial][google.cloud.aiplatform.v1.VizierService.StopTrial].
Required. The Trial's name. Format: `projects/{project}/locations/{location}/studies/{study}/trials/{trial}`
Lists the pareto-optimal Trials for multi-objective Study or the optimal Trials for single-objective Study. The definition of pareto-optimal can be checked in wiki page. https://en.wikipedia.org/wiki/Pareto_efficiency
Request message for [VizierService.ListOptimalTrials][google.cloud.aiplatform.v1.VizierService.ListOptimalTrials].
Required. The name of the Study that the optimal Trial belongs to.
Response message for [VizierService.ListOptimalTrials][google.cloud.aiplatform.v1.VizierService.ListOptimalTrials].
The pareto-optimal Trials for multiple objective Study or the optimal trial for single objective Study. The definition of pareto-optimal can be checked in wiki page. https://en.wikipedia.org/wiki/Pareto_efficiency
LINT: LEGACY_NAMES Represents a hardware accelerator type.
Used in:
Unspecified accelerator type, which means no accelerator.
Deprecated: Nvidia Tesla K80 GPU has reached end of support, see https://cloud.google.com/compute/docs/eol/k80-eol.
Nvidia Tesla P100 GPU.
Nvidia Tesla V100 GPU.
Nvidia Tesla P4 GPU.
Nvidia Tesla T4 GPU.
Nvidia Tesla A100 GPU.
Nvidia A100 80GB GPU.
Nvidia L4 GPU.
Nvidia H100 80Gb GPU.
Nvidia H100 Mega 80Gb GPU.
Nvidia H200 141Gb GPU.
Nvidia B200 GPU.
TPU v2.
TPU v3.
TPU v4.
TPU v5.
Parameters that configure the active learning pipeline. Active learning will label the data incrementally by several iterations. For every iteration, it will select a batch of data based on the sampling strategy.
Used in:
Required. Max human labeling DataItems. The rest part will be labeled by machine.
Max number of human labeled DataItems.
Max percent of total DataItems for human labeling.
Active learning data sampling config. For every active learning labeling iteration, it will select a batch of data based on the sampling strategy.
CMLE training config. For every active learning labeling iteration, system will train a machine learning model on CMLE. The trained model will be used by data sampling algorithm to select DataItems.
Used to assign specific AnnotationSpec to a particular area of a DataItem or the whole part of the DataItem.
Used in:
,Output only. Resource name of the Annotation.
Required. Google Cloud Storage URI points to a YAML file describing [payload][google.cloud.aiplatform.v1.Annotation.payload]. The schema is defined as an [OpenAPI 3.0.2 Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/, note that the chosen schema must be consistent with the parent Dataset's [metadata][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri].
Required. The schema of the payload can be found in [payload_schema][google.cloud.aiplatform.v1.Annotation.payload_schema_uri].
Output only. Timestamp when this Annotation was created.
Output only. Timestamp when this Annotation was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Output only. The source of the Annotation.
Optional. The labels with user-defined metadata to organize your Annotations. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Annotation(System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for each Annotation: * "aiplatform.googleapis.com/annotation_set_name": optional, name of the UI's annotation set this Annotation belongs to. If not set, the Annotation is not visible in the UI. * "aiplatform.googleapis.com/payload_schema": output only, its value is the [payload_schema's][google.cloud.aiplatform.v1.Annotation.payload_schema_uri] title.
The generic reusable api auth config.
Used in:
The auth config.
The API secret.
The API secret.
Used in:
, , ,Required. The SecretManager secret version resource name storing API key. e.g. projects/{project}/secrets/{secret}/versions/{version}
Instance of a general artifact.
Used as response type in: MetadataService.CreateArtifact, MetadataService.GetArtifact, MetadataService.UpdateArtifact
Used as field type in:
, , , ,Output only. The resource name of the Artifact.
User provided display name of the Artifact. May be up to 128 Unicode characters.
The uniform resource identifier of the artifact file. May be empty if there is no actual artifact file.
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Artifacts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Artifact (System labels are excluded).
Output only. Timestamp when this Artifact was created.
Output only. Timestamp when this Artifact was last updated.
The state of this Artifact. This is a property of the Artifact, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines), and the system does not prescribe or check the validity of state transitions.
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
Properties of the Artifact. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
Description of the Artifact
Describes the state of the Artifact.
Used in:
Unspecified state for the Artifact.
A state used by systems like Vertex AI Pipelines to indicate that the underlying data item represented by this Artifact is being created.
A state indicating that the Artifact should exist, unless something external to the system deletes it.
Metadata information for [NotebookService.AssignNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.AssignNotebookRuntime].
The operation generic information.
A human-readable message that shows the intermediate progress details of NotebookRuntime.
Attribution that explains a particular prediction output.
Used in:
,Output only. Model predicted output if the input instance is constructed from the baselines of all the features defined in [ExplanationMetadata.inputs][google.cloud.aiplatform.v1.ExplanationMetadata.inputs]. The field name of the output is determined by the key in [ExplanationMetadata.outputs][google.cloud.aiplatform.v1.ExplanationMetadata.outputs]. If the Model's predicted output has multiple dimensions (rank > 1), this is the value in the output located by [output_index][google.cloud.aiplatform.v1.Attribution.output_index]. If there are multiple baselines, their output values are averaged.
Output only. Model predicted output on the corresponding [explanation instance][ExplainRequest.instances]. The field name of the output is determined by the key in [ExplanationMetadata.outputs][google.cloud.aiplatform.v1.ExplanationMetadata.outputs]. If the Model predicted output has multiple dimensions, this is the value in the output located by [output_index][google.cloud.aiplatform.v1.Attribution.output_index].
Output only. Attributions of each explained feature. Features are extracted from the [prediction instances][google.cloud.aiplatform.v1.ExplainRequest.instances] according to [explanation metadata for inputs][google.cloud.aiplatform.v1.ExplanationMetadata.inputs]. The value is a struct, whose keys are the name of the feature. The values are how much the feature in the [instance][google.cloud.aiplatform.v1.ExplainRequest.instances] contributed to the predicted result. The format of the value is determined by the feature's input format: * If the feature is a scalar value, the attribution value is a [floating number][google.protobuf.Value.number_value]. * If the feature is an array of scalar values, the attribution value is an [array][google.protobuf.Value.list_value]. * If the feature is a struct, the attribution value is a [struct][google.protobuf.Value.struct_value]. The keys in the attribution value struct are the same as the keys in the feature struct. The formats of the values in the attribution struct are determined by the formats of the values in the feature struct. The [ExplanationMetadata.feature_attributions_schema_uri][google.cloud.aiplatform.v1.ExplanationMetadata.feature_attributions_schema_uri] field, pointed to by the [ExplanationSpec][google.cloud.aiplatform.v1.ExplanationSpec] field of the [Endpoint.deployed_models][google.cloud.aiplatform.v1.Endpoint.deployed_models] object, points to the schema file that describes the features and their attribution values (if it is populated).
Output only. The index that locates the explained prediction output. If the prediction output is a scalar value, output_index is not populated. If the prediction output has multiple dimensions, the length of the output_index list is the same as the number of dimensions of the output. The i-th element in output_index is the element index of the i-th dimension of the output vector. Indices start from 0.
Output only. The display name of the output identified by [output_index][google.cloud.aiplatform.v1.Attribution.output_index]. For example, the predicted class name by a multi-classification Model. This field is only populated iff the Model predicts display names as a separate field along with the explained output. The predicted display name must has the same shape of the explained output, and can be located using output_index.
Output only. Error of [feature_attributions][google.cloud.aiplatform.v1.Attribution.feature_attributions] caused by approximation used in the explanation method. Lower value means more precise attributions. * For Sampled Shapley [attribution][google.cloud.aiplatform.v1.ExplanationParameters.sampled_shapley_attribution], increasing [path_count][google.cloud.aiplatform.v1.SampledShapleyAttribution.path_count] might reduce the error. * For Integrated Gradients [attribution][google.cloud.aiplatform.v1.ExplanationParameters.integrated_gradients_attribution], increasing [step_count][google.cloud.aiplatform.v1.IntegratedGradientsAttribution.step_count] might reduce the error. * For [XRAI attribution][google.cloud.aiplatform.v1.ExplanationParameters.xrai_attribution], increasing [step_count][google.cloud.aiplatform.v1.XraiAttribution.step_count] might reduce the error. See [this introduction](/vertex-ai/docs/explainable-ai/overview) for more information.
Output only. Name of the explain output. Specified as the key in [ExplanationMetadata.outputs][google.cloud.aiplatform.v1.ExplanationMetadata.outputs].
Metadata of the backend deployed model.
Used in:
Optional. The model that the user will send the augmented prompt for content generation.
Optional. The model version of the backend deployed model.
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration. Each Model supporting these resources documents its specific guidelines.
Used in:
, , ,Immutable. The minimum number of replicas this DeployedModel will be always deployed on. If traffic against it increases, it may dynamically be deployed onto more replicas up to [max_replica_count][google.cloud.aiplatform.v1.AutomaticResources.max_replica_count], and as traffic decreases, some of these extra replicas may be freed. If the requested value is too large, the deployment will error.
Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, a no upper bound for scaling under heavy traffic will be assume, though Vertex AI may be unable to scale beyond certain replica number.
The metric specification that defines the target resource utilization (CPU utilization, accelerator's duty cycle, and so on) for calculating the desired replica count.
Used in:
Required. The resource metric name. Supported metrics: * For Online Prediction: * `aiplatform.googleapis.com/prediction/online/accelerator/duty_cycle` * `aiplatform.googleapis.com/prediction/online/cpu/utilization`
The target resource utilization in percentage (1% - 100%) for the given metric; once the real usage deviates from the target by a certain percentage, the machine replicas change. The default value is 60 (representing 60%) if not provided.
The storage details for Avro input content.
Used in:
Required. Google Cloud Storage location.
Runtime operation information for [PipelineService.BatchCancelPipelineJobs][google.cloud.aiplatform.v1.PipelineService.BatchCancelPipelineJobs].
The common part of the operation metadata.
Response message for [PipelineService.BatchCancelPipelineJobs][google.cloud.aiplatform.v1.PipelineService.BatchCancelPipelineJobs].
PipelineJobs cancelled.
Details of operations that perform batch create Features.
Operation metadata for Feature.
Request message for [FeaturestoreService.BatchCreateFeatures][google.cloud.aiplatform.v1.FeaturestoreService.BatchCreateFeatures]. Request message for [FeatureRegistryService.BatchCreateFeatures][google.cloud.aiplatform.v1.FeatureRegistryService.BatchCreateFeatures].
Used as request type in: FeatureRegistryService.BatchCreateFeatures, FeaturestoreService.BatchCreateFeatures
Required. The resource name of the EntityType/FeatureGroup to create the batch of Features under. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` `projects/{project}/locations/{location}/featureGroups/{feature_group}`
Required. The request message specifying the Features to create. All Features must be created under the same parent EntityType / FeatureGroup. The `parent` field in each child request message can be omitted. If `parent` is set in a child request, then the value must match the `parent` value in this request message.
Response message for [FeaturestoreService.BatchCreateFeatures][google.cloud.aiplatform.v1.FeaturestoreService.BatchCreateFeatures].
The Features created.
A description of resources that are used for performing batch operations, are dedicated to a Model, and need manual configuration.
Used in:
Required. Immutable. The specification of a single machine.
Immutable. The number of machine replicas used at the start of the batch operation. If not set, Vertex AI decides starting number, not greater than [max_replica_count][google.cloud.aiplatform.v1.BatchDedicatedResources.max_replica_count]
Immutable. The maximum number of machine replicas the batch operation may be scaled to. The default value is 10.
Response message for [PipelineService.BatchDeletePipelineJobs][google.cloud.aiplatform.v1.PipelineService.BatchDeletePipelineJobs].
PipelineJobs deleted.
Runtime operation information for [MigrationService.BatchMigrateResources][google.cloud.aiplatform.v1.MigrationService.BatchMigrateResources].
The common part of the operation metadata.
Partial results that reflect the latest migration operation progress.
Represents a partial result in batch migration operation for one [MigrateResourceRequest][google.cloud.aiplatform.v1.MigrateResourceRequest].
Used in:
If the resource's migration is ongoing, none of the result will be set. If the resource's migration is finished, either error or one of the migrated resource name will be filled.
The error result of the migration request in case of failure.
Migrated model resource name.
Migrated dataset resource name.
It's the same as the value in [BatchMigrateResourcesRequest.migrate_resource_requests][google.cloud.aiplatform.v1.BatchMigrateResourcesRequest.migrate_resource_requests].
Response message for [MigrationService.BatchMigrateResources][google.cloud.aiplatform.v1.MigrationService.BatchMigrateResources].
Successfully migrated resources.
A job that uses a [Model][google.cloud.aiplatform.v1.BatchPredictionJob.model] to produce predictions on multiple [input instances][google.cloud.aiplatform.v1.BatchPredictionJob.input_config]. If predictions for significant portion of the instances fail, the job may finish without attempting predictions for all remaining instances.
Used as response type in: JobService.CreateBatchPredictionJob, JobService.GetBatchPredictionJob
Used as field type in:
,Output only. Resource name of the BatchPredictionJob.
Required. The user-defined name of this BatchPredictionJob.
The name of the Model resource that produces the predictions via this job, must share the same ancestor Location. Starting this job has no impact on any existing deployments of the Model and their resources. Exactly one of model and unmanaged_container_model must be set. The model resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. The model resource could also be a publisher model. Example: `publishers/{publisher}/models/{model}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`
Output only. The version ID of the Model that produces the predictions via this job.
Contains model information necessary to perform batch prediction without requiring uploading to model registry. Exactly one of model and unmanaged_container_model must be set.
Required. Input configuration of the instances on which predictions are performed. The schema of any single instance may be specified via the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri].
Configuration for how to convert batch prediction input instances to the prediction instances that are sent to the Model.
The parameters that govern the predictions. The schema of the parameters may be specified via the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [parameters_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri].
Required. The Configuration specifying where output predictions should be written. The schema of any single prediction may be specified as a concatenation of [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] and [prediction_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.prediction_schema_uri].
The config of resources used by the Model during the batch prediction. If the Model [supports][google.cloud.aiplatform.v1.Model.supported_deployment_resources_types] DEDICATED_RESOURCES this config may be provided (and the job will use these resources), if the Model doesn't support AUTOMATIC_RESOURCES, this config must be provided.
The service account that the DeployedModel's container runs as. If not specified, a system generated one will be used, which has minimal permissions and the custom container, if used, may not have enough permission to access other Google Cloud resources. Users deploying the Model must have the `iam.serviceAccounts.actAs` permission on this service account.
Immutable. Parameters configuring the batch behavior. Currently only applicable when [dedicated_resources][google.cloud.aiplatform.v1.BatchPredictionJob.dedicated_resources] are used (in other cases Vertex AI does the tuning itself).
Generate explanation with the batch prediction results. When set to `true`, the batch prediction output changes based on the `predictions_format` field of the [BatchPredictionJob.output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config] object: * `bigquery`: output includes a column named `explanation`. The value is a struct that conforms to the [Explanation][google.cloud.aiplatform.v1.Explanation] object. * `jsonl`: The JSON objects on each line include an additional entry keyed `explanation`. The value of the entry is a JSON object that conforms to the [Explanation][google.cloud.aiplatform.v1.Explanation] object. * `csv`: Generating explanations for CSV format is not supported. If this field is set to true, either the [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] or [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] must be populated.
Explanation configuration for this BatchPredictionJob. Can be specified only if [generate_explanation][google.cloud.aiplatform.v1.BatchPredictionJob.generate_explanation] is set to `true`. This value overrides the value of [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec]. All fields of [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] are optional in the request. If a field of the [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] object is not populated, the corresponding field of the [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] object is inherited.
Output only. Information further describing the output of this job.
Output only. The detailed state of the job.
Output only. Only populated when the job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED.
Output only. Partial failures encountered. For example, single files that can't be read. This field never exceeds 20 entries. Status details fields contain standard Google Cloud error details.
Output only. Information about resources that had been consumed by this job. Provided in real time at best effort basis, as well as a final value once the job completes. Note: This field currently may be not populated for batch predictions that use AutoML Models.
Output only. Statistics on completed and failed prediction instances.
Output only. Time when the BatchPredictionJob was created.
Output only. Time when the BatchPredictionJob for the first time entered the `JOB_STATE_RUNNING` state.
Output only. Time when the BatchPredictionJob entered any of the following states: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`.
Output only. Time when the BatchPredictionJob was most recently updated.
The labels with user-defined metadata to organize BatchPredictionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Customer-managed encryption key options for a BatchPredictionJob. If this is set, then all resources created by the BatchPredictionJob will be encrypted with the provided encryption key.
For custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send `stderr` and `stdout` streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging/pricing). User can disable container logging by setting this flag to true.
Output only. Reserved for future use.
Output only. Reserved for future use.
Configures the input to [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob]. See [Model.supported_input_storage_formats][google.cloud.aiplatform.v1.Model.supported_input_storage_formats] for Model's supported input formats, and how instances should be expressed via any of them.
Used in:
Required. The source of the input.
The Cloud Storage location for the input instances.
The BigQuery location of the input table. The schema of the table should be in the format described by the given context OpenAPI Schema, if one is provided. The table may contain additional columns that are not described by the schema, and they will be ignored.
Required. The format in which instances are given, must be one of the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model] [supported_input_storage_formats][google.cloud.aiplatform.v1.Model.supported_input_storage_formats].
Configuration defining how to transform batch prediction input instances to the instances that the Model accepts.
Used in:
The format of the instance that the Model accepts. Vertex AI will convert compatible [batch prediction input instance formats][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.instances_format] to the specified format. Supported values are: * `object`: Each input is converted to JSON object format. * For `bigquery`, each row is converted to an object. * For `jsonl`, each line of the JSONL input must be an object. * Does not apply to `csv`, `file-list`, `tf-record`, or `tf-record-gzip`. * `array`: Each input is converted to JSON array format. * For `bigquery`, each row is converted to an array. The order of columns is determined by the BigQuery column order, unless [included_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.included_fields] is populated. [included_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.included_fields] must be populated for specifying field orders. * For `jsonl`, if each line of the JSONL input is an object, [included_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.included_fields] must be populated for specifying field orders. * Does not apply to `csv`, `file-list`, `tf-record`, or `tf-record-gzip`. If not specified, Vertex AI converts the batch prediction input as follows: * For `bigquery` and `csv`, the behavior is the same as `array`. The order of columns is the same as defined in the file or table, unless [included_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.included_fields] is populated. * For `jsonl`, the prediction instance format is determined by each line of the input. * For `tf-record`/`tf-record-gzip`, each record will be converted to an object in the format of `{"b64": <value>}`, where `<value>` is the Base64-encoded string of the content of the record. * For `file-list`, each file in the list will be converted to an object in the format of `{"b64": <value>}`, where `<value>` is the Base64-encoded string of the content of the file.
The name of the field that is considered as a key. The values identified by the key field is not included in the transformed instances that is sent to the Model. This is similar to specifying this name of the field in [excluded_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.excluded_fields]. In addition, the batch prediction output will not include the instances. Instead the output will only include the value of the key field, in a field named `key` in the output: * For `jsonl` output format, the output will have a `key` field instead of the `instance` field. * For `csv`/`bigquery` output format, the output will have have a `key` column instead of the instance feature columns. The input must be JSONL with objects at each line, CSV, BigQuery or TfRecord.
Fields that will be included in the prediction instance that is sent to the Model. If [instance_type][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.instance_type] is `array`, the order of field names in included_fields also determines the order of the values in the array. When included_fields is populated, [excluded_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.excluded_fields] must be empty. The input must be JSONL with objects at each line, BigQuery or TfRecord.
Fields that will be excluded in the prediction instance that is sent to the Model. Excluded will be attached to the batch prediction output if [key_field][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.key_field] is not specified. When excluded_fields is populated, [included_fields][google.cloud.aiplatform.v1.BatchPredictionJob.InstanceConfig.included_fields] must be empty. The input must be JSONL with objects at each line, BigQuery or TfRecord.
Configures the output of [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob]. See [Model.supported_output_storage_formats][google.cloud.aiplatform.v1.Model.supported_output_storage_formats] for supported output formats, and how predictions are expressed via any of them.
Used in:
Required. The destination of the output.
The Cloud Storage location of the directory where the output is to be written to. In the given directory a new directory is created. Its name is `prediction-<model-display-name>-<job-create-time>`, where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. Inside of it files `predictions_0001.<extension>`, `predictions_0002.<extension>`, ..., `predictions_N.<extension>` are created where `<extension>` depends on chosen [predictions_format][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.predictions_format], and N may equal 0001 and depends on the total number of successfully predicted instances. If the Model has both [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] and [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] schemata defined then each such file contains predictions as per the [predictions_format][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.predictions_format]. If prediction for any instance failed (partially or completely), then an additional `errors_0001.<extension>`, `errors_0002.<extension>`,..., `errors_N.<extension>` files are created (N depends on total number of failed predictions). These files contain the failed instances, as per their schema, followed by an additional `error` field which as value has [google.rpc.Status][google.rpc.Status] containing only `code` and `message` fields.
The BigQuery project or dataset location where the output is to be written to. If project is provided, a new dataset is created with name `prediction_<model-display-name>_<job-create-time>` where <model-display-name> is made BigQuery-dataset-name compatible (for example, most special characters become underscores), and timestamp is in YYYY_MM_DDThh_mm_ss_sssZ "based on ISO-8601" format. In the dataset two tables will be created, `predictions`, and `errors`. If the Model has both [instance][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] and [prediction][google.cloud.aiplatform.v1.PredictSchemata.parameters_schema_uri] schemata defined then the tables have columns as follows: The `predictions` table contains instances for which the prediction succeeded, it has columns as per a concatenation of the Model's instance and prediction schemata. The `errors` table contains rows for which the prediction has failed, it has instance columns, as per the instance schema, followed by a single "errors" column, which as values has [google.rpc.Status][google.rpc.Status] represented as a STRUCT, and containing only `code` and `message`.
Required. The format in which Vertex AI gives the predictions, must be one of the [Model's][google.cloud.aiplatform.v1.BatchPredictionJob.model] [supported_output_storage_formats][google.cloud.aiplatform.v1.Model.supported_output_storage_formats].
Further describes this job's output. Supplements [output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config].
Used in:
The output location into which prediction output is written.
Output only. The full path of the Cloud Storage directory created, into which the prediction output is written.
Output only. The path of the BigQuery dataset created, in `bq://projectId.bqDatasetId` format, into which the prediction output is written.
Output only. The name of the BigQuery table created, in `predictions_<timestamp>` format, into which the prediction output is written. Can be used by UI to generate the BigQuery output path, for example.
Details of operations that batch reads Feature values.
Operation metadata for Featurestore batch read Features values.
Selects Features of an EntityType to read values of and specifies read settings.
Used in:
Required. ID of the EntityType to select Features. The EntityType id is the [entity_type_id][google.cloud.aiplatform.v1.CreateEntityTypeRequest.entity_type_id] specified during EntityType creation.
Required. Selectors choosing which Feature values to read from the EntityType.
Per-Feature settings for the batch read.
Describe pass-through fields in read_instance source.
Used in:
Required. The name of the field in the CSV header or the name of the column in BigQuery table. The naming restriction is the same as [Feature.name][google.cloud.aiplatform.v1.Feature.name].
Response message for [FeaturestoreService.BatchReadFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.BatchReadFeatureValues].
(message has no fields)
The BigQuery location for the output content.
Used in:
, , , , ,Required. BigQuery URI to a project or table, up to 2000 characters long. When only the project is specified, the Dataset and Table is created. When the full table reference is specified, the Dataset must exist and table must not exist. Accepted forms: * BigQuery path. For example: `bq://projectId` or `bq://projectId.bqDatasetId` or `bq://projectId.bqDatasetId.bqTableId`.
The BigQuery location for the input content.
Used in:
, , , ,Required. BigQuery URI to a table, up to 2000 characters long. Accepted forms: * BigQuery path. For example: `bq://projectId.bqDatasetId.bqTableId`.
Input for bleu metric.
Used in:
Required. Spec for bleu score metric.
Required. Repeated bleu instances.
Spec for bleu instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Bleu metric value for an instance.
Used in:
Output only. Bleu score.
Results for bleu metric.
Used in:
Output only. Bleu metric values.
Spec for bleu score metric - calculates the precision of n-grams in the prediction as compared to reference - returns a score ranging between 0 to 1.
Used in:
Optional. Whether to use_effective_order to compute bleu score.
Content blob. It's preferred to send as [text][google.cloud.aiplatform.v1.Part.text] directly rather than raw bytes.
Used in:
Required. The IANA standard MIME type of the source data.
Required. Raw bytes.
Config for blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
Used in:
,The standard deviation of the blur kernel for the blurred baseline. The same blurring parameter is used for both the height and the width dimension. If not set, the method defaults to the zero (i.e. black for images) baseline.
A list of boolean values.
Used in:
A list of bool values.
A resource used in LLM queries for users to explicitly specify what to cache and how to cache.
Used as response type in: GenAiCacheService.CreateCachedContent, GenAiCacheService.GetCachedContent, GenAiCacheService.UpdateCachedContent
Used as field type in:
, ,Expiration time of the cached content.
Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
Input only. The TTL for this resource. The expiration time is computed: now + TTL.
Immutable. Identifier. The server-generated resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content}
Optional. Immutable. The user-generated meaningful display name of the cached content.
Immutable. The name of the `Model` to use for cached content. Currently, only the published Gemini base models are supported, in form of projects/{PROJECT}/locations/{LOCATION}/publishers/google/models/{MODEL}
Optional. Input only. Immutable. Developer set system instruction. Currently, text only
Optional. Input only. Immutable. The content to cache
Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
Optional. Input only. Immutable. Tool config. This config is shared for all tools
Output only. Creation time of the cache entry.
Output only. When the cache entry was last updated in UTC time.
Output only. Metadata on the usage of the cached content.
Input only. Immutable. Customer-managed encryption key spec for a `CachedContent`. If set, this `CachedContent` and all its sub-resources will be secured by this key.
Metadata on the usage of the cached content.
Used in:
Total number of tokens that the cached content consumes.
Number of text characters.
Number of images.
Duration of video in seconds.
Duration of audio in seconds.
A response candidate generated from the model.
Used in:
Output only. Index of the candidate.
Output only. Content parts of the candidate.
Output only. Confidence score of the candidate.
Output only. Average log probability score of the candidate.
Output only. Log-likelihood scores for the response tokens and top tokens
Output only. The reason why the model stopped generating tokens. If empty, the model has not stopped generating the tokens.
Output only. List of ratings for the safety of a response candidate. There is at most one rating per category.
Output only. Describes the reason the mode stopped generating tokens in more detail. This is only filled when `finish_reason` is set.
Output only. Source attribution of the generated content.
Output only. Metadata specifies sources used to ground generated content.
The reason why the model stopped generating tokens. If empty, the model has not stopped generating the tokens.
Used in:
The finish reason is unspecified.
Token generation reached a natural stopping point or a configured stop sequence.
Token generation reached the configured maximum output tokens.
Token generation stopped because the content potentially contains safety violations. NOTE: When streaming, [content][google.cloud.aiplatform.v1.Candidate.content] is empty if content filters blocks the output.
Token generation stopped because the content potentially contains copyright violations.
All other reasons that stopped the token generation.
Token generation stopped because the content contains forbidden terms.
Token generation stopped for potentially containing prohibited content.
Token generation stopped because the content potentially contains Sensitive Personally Identifiable Information (SPII).
The function call generated by the model is invalid.
This message will be placed in the metadata field of a google.longrunning.Operation associated with a CheckTrialEarlyStoppingState request.
Operation metadata for suggesting Trials.
The name of the Study that the Trial belongs to.
The Trial name.
Response message for [VizierService.CheckTrialEarlyStoppingState][google.cloud.aiplatform.v1.VizierService.CheckTrialEarlyStoppingState].
True if the Trial should stop.
Describes the machine learning model version checkpoint.
Used in:
The ID of the checkpoint.
The epoch of the checkpoint.
The step of the checkpoint.
Source attributions for content.
Used in:
Output only. Start index into the content.
Output only. End index into the content.
Output only. Url reference of the attribution.
Output only. Title of the attribution.
Output only. License of the attribution.
Output only. Publication date of the attribution.
A collection of source attributions for a piece of content.
Used in:
Output only. List of citations.
Claim that is extracted from the input text and facts that support it.
Used in:
Index in the input text where the claim starts (inclusive).
Index in the input text where the claim ends (exclusive).
Indexes of the facts supporting this claim.
Confidence score of this corroboration.
Configurations (e.g. inference timeout) that are applied on your endpoints.
Used in:
Customizable online prediction request timeout.
Result of executing the [ExecutableCode]. Always follows a `part` containing the [ExecutableCode].
Used in:
Required. Outcome of the code execution.
Optional. Contains stdout when code execution is successful, stderr or other description otherwise.
Enumeration of possible outcomes of the code execution.
Used in:
Unspecified status. This value should not be used.
Code execution completed successfully.
Code execution finished but with a failure. `stderr` should contain the reason.
Code execution ran for too long, and was cancelled. There may or may not be a partial output present.
Input for coherence metric.
Used in:
Required. Spec for coherence score metric.
Required. Coherence instance.
Spec for coherence instance.
Used in:
Required. Output of the evaluated model.
Spec for coherence result.
Used in:
Output only. Coherence score.
Output only. Explanation for coherence score.
Output only. Confidence for coherence score.
Spec for coherence score metric.
Used in:
Optional. Which version to use for evaluation.
Input for Comet metric.
Used in:
Required. Spec for comet metric.
Required. Comet instance.
Spec for Comet instance - The fields used for evaluation are dependent on the comet version.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Optional. Source text in original language.
Spec for Comet result - calculates the comet score for the given instance using the version specified in the spec.
Used in:
Output only. Comet score. Range depends on version.
Spec for Comet metric.
Used in:
Required. Which version to use for evaluation.
Optional. Source language in BCP-47 format.
Optional. Target language in BCP-47 format. Covers both prediction and reference.
Comet version options.
Used in:
Comet version unspecified.
Comet 22 for translation + source + reference (source-reference-combined).
Success and error statistics of processing multiple entities (for example, DataItems or structured data rows) in batch.
Used in:
Output only. The number of entities that had been processed successfully.
Output only. The number of entities for which any error was encountered.
Output only. In cases when enough errors are encountered a job, pipeline, or operation may be failed as a whole. Below is the number of entities for which the processing had not been finished (either in successful or failed state). Set to -1 if the number is unknown (for example, the operation failed before the total entity number could be collected).
Output only. The number of the successful forecast points that are generated by the forecasting model. This is ONLY used by the forecasting batch prediction.
The Container Registry location for the container image.
Used in:
Required. Container Registry URI of a container image. Only Google Container Registry and Artifact Registry are supported now. Accepted forms: * Google Container Registry path. For example: `gcr.io/projectId/imageName:tag`. * Artifact Registry path. For example: `us-central1-docker.pkg.dev/projectId/repoName/imageName:tag`. If a tag is not specified, "latest" will be used as the default tag.
The spec of a Container.
Used in:
Required. The URI of a container image in the Container Registry that is to be run on each worker replica.
The command to be invoked when the container is started. It overrides the entrypoint instruction in Dockerfile when provided.
The arguments to be passed when starting the container.
Environment variables to be passed to the container. Maximum limit is 100.
The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
Used in:
, , , , , , , ,Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset.
Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types.
Instance of a general context.
Used as response type in: MetadataService.CreateContext, MetadataService.GetContext, MetadataService.UpdateContext
Used as field type in:
, , ,Immutable. The resource name of the Context.
User provided display name of the Context. May be up to 128 Unicode characters.
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Contexts. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Context (System labels are excluded).
Output only. Timestamp when this Context was created.
Output only. Timestamp when this Context was last updated.
Output only. A list of resource names of Contexts that are parents of this Context. A Context may have at most 10 parent_contexts.
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
The version of the schema in schema_name to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
Properties of the Context. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
Description of the Context
Details of [ModelService.CopyModel][google.cloud.aiplatform.v1.ModelService.CopyModel] operation.
The common part of the operation metadata.
Response message of [ModelService.CopyModel][google.cloud.aiplatform.v1.ModelService.CopyModel] operation.
The name of the copied Model resource. Format: `projects/{project}/locations/{location}/models/{model}`
Output only. The version ID of the model that is copied.
RagCorpus status.
Used in:
Output only. RagCorpus life state.
Output only. Only when the `state` field is ERROR.
RagCorpus life state.
Used in:
This state is not supposed to happen.
RagCorpus resource entry is initialized, but hasn't done validation.
RagCorpus is provisioned successfully and is ready to serve.
RagCorpus is in a problematic situation. See `error_message` field for details.
Parameters that can be overrided per request.
Used in:
Optional. Only return claims with citation score larger than the threshold.
Runtime operation information for [DatasetService.CreateDataset][google.cloud.aiplatform.v1.DatasetService.CreateDataset].
The operation generic information.
Runtime operation information for [DatasetService.CreateDatasetVersion][google.cloud.aiplatform.v1.DatasetService.CreateDatasetVersion].
The common part of the operation metadata.
Runtime operation information for CreateDeploymentResourcePool method.
The operation generic information.
Runtime operation information for [EndpointService.CreateEndpoint][google.cloud.aiplatform.v1.EndpointService.CreateEndpoint].
The operation generic information.
Details of operations that perform create EntityType.
Operation metadata for EntityType.
Details of operations that perform create FeatureGroup.
Operation metadata for FeatureGroup.
Details of operations that perform create FeatureOnlineStore.
Operation metadata for FeatureOnlineStore.
Details of operations that perform create Feature.
Operation metadata for Feature.
Request message for [FeaturestoreService.CreateFeature][google.cloud.aiplatform.v1.FeaturestoreService.CreateFeature]. Request message for [FeatureRegistryService.CreateFeature][google.cloud.aiplatform.v1.FeatureRegistryService.CreateFeature].
Used as request type in: FeatureRegistryService.CreateFeature, FeaturestoreService.CreateFeature
Used as field type in:
Required. The resource name of the EntityType or FeatureGroup to create a Feature. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}`
Required. The Feature to create.
Required. The ID to use for the Feature, which will become the final component of the Feature's resource name. This value may be up to 128 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within an EntityType/FeatureGroup.
Details of operations that perform create FeatureView.
Operation metadata for FeatureView Create.
Details of operations that perform create Featurestore.
Operation metadata for Featurestore.
Runtime operation information for [IndexEndpointService.CreateIndexEndpoint][google.cloud.aiplatform.v1.IndexEndpointService.CreateIndexEndpoint].
The operation generic information.
Runtime operation information for [IndexService.CreateIndex][google.cloud.aiplatform.v1.IndexService.CreateIndex].
The operation generic information.
The operation metadata with regard to Matching Engine Index operation.
Details of operations that perform [MetadataService.CreateMetadataStore][google.cloud.aiplatform.v1.MetadataService.CreateMetadataStore].
Operation metadata for creating a MetadataStore.
Metadata information for [NotebookService.CreateNotebookExecutionJob][google.cloud.aiplatform.v1.NotebookService.CreateNotebookExecutionJob].
The operation generic information.
A human-readable message that shows the intermediate progress details of NotebookRuntime.
Request message for [NotebookService.CreateNotebookExecutionJob]
Used as request type in: NotebookService.CreateNotebookExecutionJob
Used as field type in:
Required. The resource name of the Location to create the NotebookExecutionJob. Format: `projects/{project}/locations/{location}`
Required. The NotebookExecutionJob to create.
Optional. User specified ID for the NotebookExecutionJob.
Metadata information for [NotebookService.CreateNotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookService.CreateNotebookRuntimeTemplate].
The operation generic information.
Details of operations that perform create PersistentResource.
Operation metadata for PersistentResource.
Progress Message for Create LRO
Request message for [PipelineService.CreatePipelineJob][google.cloud.aiplatform.v1.PipelineService.CreatePipelineJob].
Used as request type in: PipelineService.CreatePipelineJob
Used as field type in:
Required. The resource name of the Location to create the PipelineJob in. Format: `projects/{project}/locations/{location}`
Required. The PipelineJob to create.
The ID to use for the PipelineJob, which will become the final component of the PipelineJob name. If not provided, an ID will be automatically generated. This value should be less than 128 characters, and valid characters are `/[a-z][0-9]-/`.
Runtime operation information for [VertexRagDataService.CreateRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.CreateRagCorpus].
The operation generic information.
Details of [ReasoningEngineService.CreateReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.CreateReasoningEngine] operation.
The common part of the operation metadata.
Details of operations that perform create FeatureGroup.
Operation metadata for Feature.
Runtime operation information for [SpecialistPoolService.CreateSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.CreateSpecialistPool].
The operation generic information.
Details of operations that perform create Tensorboard.
Operation metadata for Tensorboard.
Request message for [TensorboardService.CreateTensorboardRun][google.cloud.aiplatform.v1.TensorboardService.CreateTensorboardRun].
Used as request type in: TensorboardService.CreateTensorboardRun
Used as field type in:
Required. The resource name of the TensorboardExperiment to create the TensorboardRun in. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
Required. The TensorboardRun to create.
Required. The ID to use for the Tensorboard run, which becomes the final component of the Tensorboard run's resource name. This value should be 1-128 characters, and valid characters are `/[a-z][0-9]-/`.
Request message for [TensorboardService.CreateTensorboardTimeSeries][google.cloud.aiplatform.v1.TensorboardService.CreateTensorboardTimeSeries].
Used as request type in: TensorboardService.CreateTensorboardTimeSeries
Used as field type in:
Required. The resource name of the TensorboardRun to create the TensorboardTimeSeries in. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Optional. The user specified unique ID to use for the TensorboardTimeSeries, which becomes the final component of the TensorboardTimeSeries's resource name. This value should match "[a-z0-9][a-z0-9-]{0, 127}"
Required. The TensorboardTimeSeries to create.
The storage details for CSV output content.
Used in:
Required. Google Cloud Storage location.
The storage details for CSV input content.
Used in:
, ,Required. Google Cloud Storage location.
Represents a job that runs custom workloads such as a Docker container or a Python package. A CustomJob can have multiple worker pools and each worker pool can have its own machine and input spec. A CustomJob will be cleaned up once the job enters terminal state (failed or succeeded).
Used as response type in: JobService.CreateCustomJob, JobService.GetCustomJob
Used as field type in:
,Output only. Resource name of a CustomJob.
Required. The display name of the CustomJob. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Required. Job spec.
Output only. The detailed state of the job.
Output only. Time when the CustomJob was created.
Output only. Time when the CustomJob for the first time entered the `JOB_STATE_RUNNING` state.
Output only. Time when the CustomJob entered any of the following states: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`.
Output only. Time when the CustomJob was most recently updated.
Output only. Only populated when job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
The labels with user-defined metadata to organize CustomJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Customer-managed encryption key options for a CustomJob. If this is set, then all resources created by the CustomJob will be encrypted with the provided encryption key.
Output only. URIs for accessing [interactive shells](https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell) (one URI for each training node). Only available if [job_spec.enable_web_access][google.cloud.aiplatform.v1.CustomJobSpec.enable_web_access] is `true`. The keys are names of each node in the training job; for example, `workerpool0-0` for the primary node, `workerpool1-0` for the first node in the second worker pool, and `workerpool1-1` for the second node in the second worker pool. The values are the URIs for each node's interactive shell.
Output only. Reserved for future use.
Output only. Reserved for future use.
Represents the spec of a CustomJob.
Used in:
, , ,Optional. The ID of the PersistentResource in the same Project and Location which to run If this is specified, the job will be run on existing machines held by the PersistentResource instead of on-demand short-live machines. The network and CMEK configs on the job should be consistent with those on the PersistentResource, otherwise, the job will be rejected.
Required. The spec of the worker pools including machine type and Docker image. All worker pools except the first one are optional and can be skipped by providing an empty value.
Scheduling options for a CustomJob.
Specifies the service account for workload run-as account. Users submitting jobs must have act-as permission on this run-as account. If unspecified, the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) for the CustomJob's project is used.
Optional. The full name of the Compute Engine [network](/compute/docs/networks-and-firewalls#networks) to which the Job should be peered. For example, `projects/12345/global/networks/myVPC`. [Format](/compute/docs/reference/rest/v1/networks/insert) is of the form `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in `12345`, and {network} is a network name. To specify this field, you must have already [configured VPC Network Peering for Vertex AI](https://cloud.google.com/vertex-ai/docs/general/vpc-peering). If this field is left unspecified, the job is not peered with any network.
Optional. A list of names for the reserved ip ranges under the VPC network that can be used for this job. If set, we will deploy the job within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
The Cloud Storage location to store the output of this CustomJob or HyperparameterTuningJob. For HyperparameterTuningJob, the baseOutputDirectory of each child CustomJob backing a Trial is set to a subdirectory of name [id][google.cloud.aiplatform.v1.Trial.id] under its parent HyperparameterTuningJob's baseOutputDirectory. The following Vertex AI environment variables will be passed to containers or python modules when this field is set: For CustomJob: * AIP_MODEL_DIR = `<base_output_directory>/model/` * AIP_CHECKPOINT_DIR = `<base_output_directory>/checkpoints/` * AIP_TENSORBOARD_LOG_DIR = `<base_output_directory>/logs/` For CustomJob backing a Trial of HyperparameterTuningJob: * AIP_MODEL_DIR = `<base_output_directory>/<trial_id>/model/` * AIP_CHECKPOINT_DIR = `<base_output_directory>/<trial_id>/checkpoints/` * AIP_TENSORBOARD_LOG_DIR = `<base_output_directory>/<trial_id>/logs/`
The ID of the location to store protected artifacts. e.g. us-central1. Populate only when the location is different than CustomJob location. List of supported locations: https://cloud.google.com/vertex-ai/docs/general/locations
Optional. The name of a Vertex AI [Tensorboard][google.cloud.aiplatform.v1.Tensorboard] resource to which this CustomJob will upload Tensorboard logs. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Optional. Whether you want Vertex AI to enable [interactive shell access](https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell) to training containers. If set to `true`, you can access interactive shells at the URIs given by [CustomJob.web_access_uris][google.cloud.aiplatform.v1.CustomJob.web_access_uris] or [Trial.web_access_uris][google.cloud.aiplatform.v1.Trial.web_access_uris] (within [HyperparameterTuningJob.trials][google.cloud.aiplatform.v1.HyperparameterTuningJob.trials]).
Optional. Whether you want Vertex AI to enable access to the customized dashboard in training chief container. If set to `true`, you can access the dashboard at the URIs given by [CustomJob.web_access_uris][google.cloud.aiplatform.v1.CustomJob.web_access_uris] or [Trial.web_access_uris][google.cloud.aiplatform.v1.Trial.web_access_uris] (within [HyperparameterTuningJob.trials][google.cloud.aiplatform.v1.HyperparameterTuningJob.trials]).
Optional. The Experiment associated with this job. Format: `projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}`
Optional. The Experiment Run associated with this job. Format: `projects/{project}/locations/{location}/metadataStores/{metadataStores}/contexts/{experiment-name}-{experiment-run-name}`
Optional. The name of the Model resources for which to generate a mapping to artifact URIs. Applicable only to some of the Google-provided custom jobs. Format: `projects/{project}/locations/{location}/models/{model}` In order to retrieve a specific version of the model, also provide the version ID or version alias. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` If no version ID or alias is specified, the "default" version will be returned. The "default" version alias is created for the first version of the model, and can be moved to other versions later on. There will be exactly one default version.
A piece of data in a Dataset. Could be an image, a video, a document or plain text.
Used in:
,Output only. The resource name of the DataItem.
Output only. Timestamp when this DataItem was created.
Output only. Timestamp when this DataItem was last updated.
Optional. The labels with user-defined metadata to organize your DataItems. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one DataItem(System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Required. The data that the DataItem represents (for example, an image or a text snippet). The schema of the payload is stored in the parent Dataset's [metadata schema's][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri] dataItemSchemaUri field.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Output only. Reserved for future use.
Output only. Reserved for future use.
A container for a single DataItem and Annotations on it.
Used in:
The DataItem.
The Annotations on the DataItem. If too many Annotations should be returned for the DataItem, this field will be truncated per annotations_limit in request. If it was, then the has_truncated_annotations will be set to true.
True if and only if the Annotations field has been truncated. It happens if more Annotations for this DataItem met the request's annotation_filter than are allowed to be returned by annotations_limit. Note that if Annotations field is not being returned due to field mask, then this field will not be set to true no matter how many Annotations are there.
DataLabelingJob is used to trigger a human labeling job on unlabeled data from the following Dataset:
Used as response type in: JobService.CreateDataLabelingJob, JobService.GetDataLabelingJob
Used as field type in:
,Output only. Resource name of the DataLabelingJob.
Required. The user-defined name of the DataLabelingJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. Display name of a DataLabelingJob.
Required. Dataset resource names. Right now we only support labeling from a single Dataset. Format: `projects/{project}/locations/{location}/datasets/{dataset}`
Labels to assign to annotations generated by this DataLabelingJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Required. Number of labelers to work on each DataItem.
Required. The Google Cloud Storage location of the instruction pdf. This pdf is shared with labelers, and provides detailed description on how to label DataItems in Datasets.
Required. Points to a YAML file stored on Google Cloud Storage describing the config for a specific type of DataLabelingJob. The schema files that can be used here are found in the https://storage.googleapis.com/google-cloud-aiplatform bucket in the /schema/datalabelingjob/inputs/ folder.
Required. Input config parameters for the DataLabelingJob.
Output only. The detailed state of the job.
Output only. Current labeling job progress percentage scaled in interval [0, 100], indicating the percentage of DataItems that has been finished.
Output only. Estimated cost(in US dollars) that the DataLabelingJob has incurred to date.
Output only. Timestamp when this DataLabelingJob was created.
Output only. Timestamp when this DataLabelingJob was updated most recently.
Output only. DataLabelingJob errors. It is only populated when job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
The labels with user-defined metadata to organize your DataLabelingJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for each DataLabelingJob: * "aiplatform.googleapis.com/schema": output only, its value is the [inputs_schema][google.cloud.aiplatform.v1.DataLabelingJob.inputs_schema_uri]'s title.
The SpecialistPools' resource names associated with this job.
Customer-managed encryption key spec for a DataLabelingJob. If set, this DataLabelingJob will be secured by this key. Note: Annotations created in the DataLabelingJob are associated with the EncryptionSpec of the Dataset they are exported to.
Parameters that configure the active learning pipeline. Active learning will label the data incrementally via several iterations. For every iteration, it will select a batch of data based on the sampling strategy.
A collection of DataItems and Annotations on them.
Used as response type in: DatasetService.GetDataset, DatasetService.UpdateDataset
Used as field type in:
, ,Output only. Identifier. The resource name of the Dataset.
Required. The user-defined name of the Dataset. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the Dataset.
Required. Points to a YAML file stored on Google Cloud Storage describing additional information about the Dataset. The schema is defined as an OpenAPI 3.0.2 Schema Object. The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/metadata/.
Required. Additional information about the Dataset.
Output only. The number of DataItems in this Dataset. Only apply for non-structured Dataset.
Output only. Timestamp when this Dataset was created.
Output only. Timestamp when this Dataset was last updated.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Datasets. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Dataset (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for each Dataset: * "aiplatform.googleapis.com/dataset_metadata_schema": output only, its value is the [metadata_schema's][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri] title.
All SavedQueries belong to the Dataset will be returned in List/Get Dataset response. The annotation_specs field will not be populated except for UI cases which will only use [annotation_spec_count][google.cloud.aiplatform.v1.SavedQuery.annotation_spec_count]. In CreateDataset request, a SavedQuery is created together if this field is set, up to one SavedQuery can be set in CreateDatasetRequest. The SavedQuery should not contain any AnnotationSpec.
Customer-managed encryption key spec for a Dataset. If set, this Dataset and all sub-resources of this Dataset will be secured by this key.
Output only. The resource name of the Artifact that was created in MetadataStore when creating the Dataset. The Artifact resource name pattern is `projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}`.
Optional. Reference to the public base model last used by the dataset. Only set for prompt datasets.
Output only. Reserved for future use.
Output only. Reserved for future use.
Describes the dataset version.
Used as response type in: DatasetService.GetDatasetVersion, DatasetService.UpdateDatasetVersion
Used as field type in:
, ,Output only. Identifier. The resource name of the DatasetVersion.
Output only. Timestamp when this DatasetVersion was created.
Output only. Timestamp when this DatasetVersion was last updated.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Output only. Name of the associated BigQuery dataset.
The user-defined name of the DatasetVersion. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Required. Output only. Additional information about the DatasetVersion.
Output only. Reference to the public base model last used by the dataset version. Only set for prompt dataset versions.
Output only. Reserved for future use.
Output only. Reserved for future use.
A description of resources that are dedicated to a DeployedModel, and that need a higher degree of manual configuration.
Used in:
, , ,Required. Immutable. The specification of a single machine used by the prediction.
Required. Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed.
Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use [min_replica_count][google.cloud.aiplatform.v1.DedicatedResources.min_replica_count] as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type).
Optional. Number of required available replicas for the deployment to succeed. This field is only needed when partial model deployment/mutation is desired. If set, the model deploy/mutate operation will succeed once available_replica_count reaches required_replica_count, and the rest of the replicas will be retried. If not set, the default required_replica_count will be min_replica_count.
Immutable. The metric specifications that overrides a resource utilization metric (CPU utilization, accelerator's duty cycle, and so on) target value (default to 60 if not set). At most one entry is allowed per metric. If [machine_spec.accelerator_count][google.cloud.aiplatform.v1.MachineSpec.accelerator_count] is above 0, the autoscaling will be based on both CPU utilization and accelerator's duty cycle metrics and scale up when either metrics exceeds its target value while scale down if both metrics are under their target value. The default target value is 60 for both metrics. If [machine_spec.accelerator_count][google.cloud.aiplatform.v1.MachineSpec.accelerator_count] is 0, the autoscaling will be based on CPU utilization metric only with default target value 60 if not explicitly set. For example, in the case of Online Prediction, if you want to override target CPU utilization to 80, you should set [autoscaling_metric_specs.metric_name][google.cloud.aiplatform.v1.AutoscalingMetricSpec.metric_name] to `aiplatform.googleapis.com/prediction/online/cpu/utilization` and [autoscaling_metric_specs.target][google.cloud.aiplatform.v1.AutoscalingMetricSpec.target] to `80`.
Optional. If true, schedule the deployment workload on [spot VMs](https://cloud.google.com/kubernetes-engine/docs/concepts/spot-vms).
Request message for [FeaturestoreService.DeleteFeature][google.cloud.aiplatform.v1.FeaturestoreService.DeleteFeature]. Request message for [FeatureRegistryService.DeleteFeature][google.cloud.aiplatform.v1.FeatureRegistryService.DeleteFeature].
Used as request type in: FeatureRegistryService.DeleteFeature, FeaturestoreService.DeleteFeature
Required. The name of the Features to be deleted. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}/features/{feature}` `projects/{project}/locations/{location}/featureGroups/{feature_group}/features/{feature}`
Details of operations that delete Feature values.
Operation metadata for Featurestore delete Features values.
Message to select entity. If an entity id is selected, all the feature values corresponding to the entity id will be deleted, including the entityId.
Used in:
Required. Selectors choosing feature values of which entity id to be deleted from the EntityType.
Message to select time range and feature. Values of the selected feature generated within an inclusive time range will be deleted. Using this option permanently deletes the feature values from the specified feature IDs within the specified time range. This might include data from the online storage. If you want to retain any deleted historical data in the online storage, you must re-ingest it.
Used in:
Required. Select feature generated within a half-inclusive time range. The time range is lower inclusive and upper exclusive.
Required. Selectors choosing which feature values to be deleted from the EntityType.
If set, data will not be deleted from online storage. When time range is older than the data in online storage, setting this to be true will make the deletion have no impact on online serving.
Response message for [FeaturestoreService.DeleteFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.DeleteFeatureValues].
Response based on which delete option is specified in the request
Response for request specifying the entities to delete
Response for request specifying time range and feature
Response message if the request uses the SelectEntity option.
Used in:
The count of deleted entity rows in the offline storage. Each row corresponds to the combination of an entity ID and a timestamp. One entity ID can have multiple rows in the offline storage.
The count of deleted entities in the online storage. Each entity ID corresponds to one entity.
Response message if the request uses the SelectTimeRangeAndFeature option.
Used in:
The count of the features or columns impacted. This is the same as the feature count in the request.
The count of modified entity rows in the offline storage. Each row corresponds to the combination of an entity ID and a timestamp. One entity ID can have multiple rows in the offline storage. Within each row, only the features specified in the request are deleted.
The count of modified entities in the online storage. Each entity ID corresponds to one entity. Within each entity, only the features specified in the request are deleted.
Details of operations that perform [MetadataService.DeleteMetadataStore][google.cloud.aiplatform.v1.MetadataService.DeleteMetadataStore].
Operation metadata for deleting a MetadataStore.
Details of operations that perform deletes of any entities.
The common part of the operation metadata.
Runtime operation information for [IndexEndpointService.DeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.DeployIndex].
The operation generic information.
The unique index id specified by user
Response message for [IndexEndpointService.DeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.DeployIndex].
The DeployedIndex that had been deployed in the IndexEndpoint.
Runtime operation information for [EndpointService.DeployModel][google.cloud.aiplatform.v1.EndpointService.DeployModel].
The operation generic information.
Response message for [EndpointService.DeployModel][google.cloud.aiplatform.v1.EndpointService.DeployModel].
The DeployedModel that had been deployed in the Endpoint.
A deployment of an Index. IndexEndpoints contain one or more DeployedIndexes.
Used in:
, , , ,Required. The user specified ID of the DeployedIndex. The ID can be up to 128 characters long and must start with a letter and only contain letters, numbers, and underscores. The ID must be unique within the project it is created in.
Required. The name of the Index this is the deployment of. We may refer to this Index as the DeployedIndex's "original" Index.
The display name of the DeployedIndex. If not provided upon creation, the Index's display_name is used.
Output only. Timestamp when the DeployedIndex was created.
Output only. Provides paths for users to send requests directly to the deployed index services running on Cloud via private services access. This field is populated if [network][google.cloud.aiplatform.v1.IndexEndpoint.network] is configured.
Output only. The DeployedIndex may depend on various data on its original Index. Additionally when certain changes to the original Index are being done (e.g. when what the Index contains is being changed) the DeployedIndex may be asynchronously updated in the background to reflect these changes. If this timestamp's value is at least the [Index.update_time][google.cloud.aiplatform.v1.Index.update_time] of the original Index, it means that this DeployedIndex and the original Index are in sync. If this timestamp is older, then to see which updates this DeployedIndex already contains (and which it does not), one must [list][google.longrunning.Operations.ListOperations] the operations that are running on the original Index. Only the successfully completed Operations with [update_time][google.cloud.aiplatform.v1.GenericOperationMetadata.update_time] equal or before this sync time are contained in this DeployedIndex.
Optional. A description of resources that the DeployedIndex uses, which to large degree are decided by Vertex AI, and optionally allows only a modest additional configuration. If min_replica_count is not set, the default value is 2 (we don't provide SLA when min_replica_count=1). If max_replica_count is not set, the default value is min_replica_count. The max allowed replica count is 1000.
Optional. A description of resources that are dedicated to the DeployedIndex, and that need a higher degree of manual configuration. The field min_replica_count must be set to a value strictly greater than 0, or else validation will fail. We don't provide SLA when min_replica_count=1. If max_replica_count is not set, the default value is min_replica_count. The max allowed replica count is 1000. Available machine types for SMALL shard: e2-standard-2 and all machine types available for MEDIUM and LARGE shard. Available machine types for MEDIUM shard: e2-standard-16 and all machine types available for LARGE shard. Available machine types for LARGE shard: e2-highmem-16, n2d-standard-32. n1-standard-16 and n1-standard-32 are still available, but we recommend e2-standard-16 and e2-highmem-16 for cost efficiency.
Optional. If true, private endpoint's access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each MatchRequest. Note that logs may incur a cost, especially if the deployed index receives a high queries per second rate (QPS). Estimate your costs before enabling this option.
Optional. If set, the authentication is enabled for the private endpoint.
Optional. A list of reserved ip ranges under the VPC network that can be used for this DeployedIndex. If set, we will deploy the index within the provided ip ranges. Otherwise, the index might be deployed to any ip ranges under the provided VPC network. The value should be the name of the address (https://cloud.google.com/compute/docs/reference/rest/v1/addresses) Example: ['vertex-ai-ip-range']. For more information about subnets and network IP ranges, please see https://cloud.google.com/vpc/docs/subnets#manually_created_subnet_ip_ranges.
Optional. The deployment group can be no longer than 64 characters (eg: 'test', 'prod'). If not set, we will use the 'default' deployment group. Creating `deployment_groups` with `reserved_ip_ranges` is a recommended practice when the peered network has multiple peering ranges. This creates your deployments from predictable IP spaces for easier traffic administration. Also, one deployment_group (except 'default') can only be used with the same reserved_ip_ranges which means if the deployment_group has been used with reserved_ip_ranges: [a, b, c], using it with [a, b] or [d, e] is disallowed. Note: we only support up to 5 deployment groups(not including 'default').
Optional. If set for PSC deployed index, PSC connection will be automatically created after deployment is done and the endpoint information is populated in private_endpoints.psc_automated_endpoints.
Used to set up the auth on the DeployedIndex's private endpoint.
Used in:
Defines the authentication provider that the DeployedIndex uses.
Configuration for an authentication provider, including support for [JSON Web Token (JWT)](https://tools.ietf.org/html/draft-ietf-oauth-json-web-token-32).
Used in:
The list of JWT [audiences](https://tools.ietf.org/html/draft-ietf-oauth-json-web-token-32#section-4.1.3). that are allowed to access. A JWT containing any of these audiences will be accepted.
A list of allowed JWT issuers. Each entry must be a valid Google service account, in the following format: `service-account-name@project-id.iam.gserviceaccount.com`
Points to a DeployedIndex.
Used in:
Immutable. A resource name of the IndexEndpoint.
Immutable. The ID of the DeployedIndex in the above IndexEndpoint.
Output only. The display name of the DeployedIndex.
A deployment of a Model. Endpoints contain one or more DeployedModels.
Used in:
, , , , ,The prediction (for example, the machine) resources that the DeployedModel uses. The user is billed for the resources (at least their minimal amount) even if the DeployedModel receives no traffic. Not all Models support all resources types. See [Model.supported_deployment_resources_types][google.cloud.aiplatform.v1.Model.supported_deployment_resources_types]. Required except for Large Model Deploy use cases.
A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
The resource name of the shared DeploymentResourcePool to deploy on. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`.
Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed.
Output only. The version ID of the model that is deployed.
The display name of the DeployedModel. If not provided upon creation, the Model's display_name is used.
Output only. Timestamp when the DeployedModel was created.
Explanation configuration for this DeployedModel. When deploying a Model using [EndpointService.DeployModel][google.cloud.aiplatform.v1.EndpointService.DeployModel], this value overrides the value of [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec]. All fields of [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] are optional in the request. If a field of [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] is not populated, the value of the same field of [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] is inherited. If the corresponding [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] is not populated, all fields of the [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] will be used for the explanation configuration.
If true, deploy the model without explainable feature, regardless the existence of [Model.explanation_spec][google.cloud.aiplatform.v1.Model.explanation_spec] or [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec].
The service account that the DeployedModel's container runs as. Specify the email address of the service account. If this service account is not specified, the container runs as a service account that doesn't have access to the resource project. Users deploying the Model must have the `iam.serviceAccounts.actAs` permission on this service account.
For custom-trained Models and AutoML Tabular Models, the container of the DeployedModel instances will send `stderr` and `stdout` streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging/pricing). User can disable container logging by setting this flag to true.
If true, online prediction access logs are sent to Cloud Logging. These logs are like standard server access logs, containing information like timestamp and latency for each prediction request. Note that logs may incur a cost, especially if your project receives prediction requests at a high queries per second rate (QPS). Estimate your costs before enabling this option.
Output only. Provide paths for users to send predict/explain/health requests directly to the deployed model services running on Cloud via private services access. This field is populated if [network][google.cloud.aiplatform.v1.Endpoint.network] is configured.
Configuration for faster model deployment.
Output only. Runtime status of the deployed model.
System labels to apply to Model Garden deployments. System labels are managed by Google for internal use only.
Optional. Spec for configuring speculative decoding.
Runtime status of the deployed model.
Used in:
Output only. The latest deployed model's status message (if any).
Output only. The time at which the status was last updated.
Output only. The number of available replicas of the deployed model.
Points to a DeployedModel.
Used in:
,Immutable. A resource name of an Endpoint.
Immutable. An ID of a DeployedModel in the above Endpoint.
A description of resources that can be shared by multiple DeployedModels, whose underlying specification consists of a DedicatedResources.
Used as response type in: DeploymentResourcePoolService.GetDeploymentResourcePool
Used as field type in:
, ,Immutable. The resource name of the DeploymentResourcePool. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
Required. The underlying DedicatedResources that the DeploymentResourcePool uses.
Customer-managed encryption key spec for a DeploymentResourcePool. If set, this DeploymentResourcePool will be secured by this key. Endpoints and the DeploymentResourcePool they deploy in need to have the same EncryptionSpec.
The service account that the DeploymentResourcePool's container(s) run as. Specify the email address of the service account. If this service account is not specified, the container(s) run as a service account that doesn't have access to the resource project. Users deploying the Models to this DeploymentResourcePool must have the `iam.serviceAccounts.actAs` permission on this service account.
If the DeploymentResourcePool is deployed with custom-trained Models or AutoML Tabular Models, the container(s) of the DeploymentResourcePool will send `stderr` and `stdout` streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging/pricing). User can disable container logging by setting this flag to true.
Output only. Timestamp when this DeploymentResourcePool was created.
Output only. Reserved for future use.
Output only. Reserved for future use.
Used in:
,Required. The ID of the Feature to apply the setting to.
Specify the field name in the export destination. If not specified, Feature ID is used.
The input content is encapsulated and uploaded in the request.
Used in:
(message has no fields)
Represents the spec of disk options.
Used in:
,Type of the boot disk (default is "pd-ssd"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) or "pd-standard" (Persistent Disk Hard Disk Drive).
Size in GB of the boot disk (default is 100GB).
A list of double values.
Used in:
A list of double values.
Describes the options to customize dynamic retrieval.
Used in:
The mode of the predictor to be used in dynamic retrieval.
Optional. The threshold to be used in dynamic retrieval. If not set, a system default value is used.
The mode of the predictor to be used in dynamic retrieval.
Used in:
Always trigger retrieval.
Run retrieval only when system decides it is necessary.
Represents a customer-managed encryption key spec that can be applied to a top-level resource.
Used in:
, , , , , , , , , , , , , , , , , , , , , , , ,Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created.
Models are deployed into it, and afterwards Endpoint is called to obtain predictions and explanations.
Used as response type in: EndpointService.GetEndpoint, EndpointService.UpdateEndpoint
Used as field type in:
, , ,Output only. The resource name of the Endpoint.
Required. The display name of the Endpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the Endpoint.
Output only. The models deployed in this Endpoint. To add or remove DeployedModels use [EndpointService.DeployModel][google.cloud.aiplatform.v1.EndpointService.DeployModel] and [EndpointService.UndeployModel][google.cloud.aiplatform.v1.EndpointService.UndeployModel] respectively.
A map from a DeployedModel's ID to the percentage of this Endpoint's traffic that should be forwarded to that DeployedModel. If a DeployedModel's ID is not listed in this map, then it receives no traffic. The traffic percentage values must add up to 100, or map must be empty if the Endpoint is to not accept any traffic at a moment.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Endpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Output only. Timestamp when this Endpoint was created.
Output only. Timestamp when this Endpoint was last updated.
Customer-managed encryption key spec for an Endpoint. If set, this Endpoint and all sub-resources of this Endpoint will be secured by this key.
Optional. The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) to which the Endpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. Only one of the fields, [network][google.cloud.aiplatform.v1.Endpoint.network] or [enable_private_service_connect][google.cloud.aiplatform.v1.Endpoint.enable_private_service_connect], can be set. [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where `{project}` is a project number, as in `12345`, and `{network}` is network name.
Deprecated: If true, expose the Endpoint via private service connect. Only one of the fields, [network][google.cloud.aiplatform.v1.Endpoint.network] or [enable_private_service_connect][google.cloud.aiplatform.v1.Endpoint.enable_private_service_connect], can be set.
Optional. Configuration for private service connect. [network][google.cloud.aiplatform.v1.Endpoint.network] and [private_service_connect_config][google.cloud.aiplatform.v1.Endpoint.private_service_connect_config] are mutually exclusive.
Output only. Resource name of the Model Monitoring job associated with this Endpoint if monitoring is enabled by [JobService.CreateModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.CreateModelDeploymentMonitoringJob]. Format: `projects/{project}/locations/{location}/modelDeploymentMonitoringJobs/{model_deployment_monitoring_job}`
Configures the request-response logging for online prediction.
If true, the endpoint will be exposed through a dedicated DNS [Endpoint.dedicated_endpoint_dns]. Your request to the dedicated DNS will be isolated from other users' traffic and will have better performance and reliability. Note: Once you enabled dedicated endpoint, you won't be able to send request to the shared DNS {region}-aiplatform.googleapis.com. The limitation will be removed soon.
Output only. DNS of the dedicated endpoint. Will only be populated if dedicated_endpoint_enabled is true. Format: `https://{endpoint_id}.{region}-{project_number}.prediction.vertexai.goog`.
Configurations that are applied to the endpoint for online prediction.
Output only. Reserved for future use.
Output only. Reserved for future use.
Tool to search public web data, powered by Vertex AI Search and Sec4 compliance.
Used in:
(message has no fields)
Selector for entityId. Getting ids from the given source.
Used in:
Details about the source data, including the location of the storage and the format.
Source of Csv
Source column that holds entity IDs. If not provided, entity IDs are extracted from the column named entity_id.
An entity type is a type of object in a system that needs to be modeled and have stored information about. For example, driver is an entity type, and driver0 is an instance of an entity type driver.
Used as response type in: FeaturestoreService.GetEntityType, FeaturestoreService.UpdateEntityType
Used as field type in:
, ,Immutable. Name of the EntityType. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` The last part entity_type is assigned by the client. The entity_type can be up to 64 characters long and can consist only of ASCII Latin letters A-Z and a-z and underscore(_), and ASCII digits 0-9 starting with a letter. The value will be unique given a featurestore.
Optional. Description of the EntityType.
Output only. Timestamp when this EntityType was created.
Output only. Timestamp when this EntityType was most recently updated.
Optional. The labels with user-defined metadata to organize your EntityTypes. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one EntityType (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Optional. Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. The default monitoring configuration for all Features with value type ([Feature.ValueType][google.cloud.aiplatform.v1.Feature.ValueType]) BOOL, STRING, DOUBLE or INT64 under this EntityType. If this is populated with [FeaturestoreMonitoringConfig.monitoring_interval] specified, snapshot analysis monitoring is enabled. Otherwise, snapshot analysis monitoring is disabled.
Optional. Config for data retention policy in offline storage. TTL in days for feature values that will be stored in offline storage. The Feature Store offline storage periodically removes obsolete feature values older than `offline_storage_ttl_days` since the feature generation time. If unset (or explicitly set to 0), default to 4000 days TTL.
Output only. Reserved for future use.
Output only. Reserved for future use.
Represents an environment variable present in a Container or Python Module.
Used in:
, , , ,Required. Name of the environment variable. Must be a valid C identifier.
Required. Variables that reference a $(VAR_NAME) are expanded using the previous defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. The $(VAR_NAME) syntax can be escaped with a double $$, ie: $$(VAR_NAME). Escaped references will never be expanded, regardless of whether the variable exists or not.
Model error analysis for each annotation.
Used in:
Attributed items for a given annotation, typically representing neighbors from the training sets constrained by the query type.
The query type used for finding the attributed items.
The outlier score of this annotated item. Usually defined as the min of all distances from attributed items.
The threshold used to determine if this annotation is an outlier or not.
Attributed items for a given annotation, typically representing neighbors from the training sets constrained by the query type.
Used in:
The unique ID for each annotation. Used by FE to allocate the annotation in DB.
The distance of this item to the annotation.
The query type used for finding the attributed items.
Used in:
Unspecified query type for model error analysis.
Query similar samples across all classes in the dataset.
Query similar samples from the same class of the input sample.
Query dissimilar samples from the same class of the input sample.
True positive, false positive, or false negative. EvaluatedAnnotation is only available under ModelEvaluationSlice with slice of `annotationSpec` dimension.
Used in:
Output only. Type of the EvaluatedAnnotation.
Output only. The model predicted annotations. For true positive, there is one and only one prediction, which matches the only one ground truth annotation in [ground_truths][google.cloud.aiplatform.v1.EvaluatedAnnotation.ground_truths]. For false positive, there is one and only one prediction, which doesn't match any ground truth annotation of the corresponding [data_item_view_id][google.cloud.aiplatform.v1.EvaluatedAnnotation.evaluated_data_item_view_id]. For false negative, there are zero or more predictions which are similar to the only ground truth annotation in [ground_truths][google.cloud.aiplatform.v1.EvaluatedAnnotation.ground_truths] but not enough for a match. The schema of the prediction is stored in [ModelEvaluation.annotation_schema_uri][google.cloud.aiplatform.v1.ModelEvaluation.annotation_schema_uri]
Output only. The ground truth Annotations, i.e. the Annotations that exist in the test data the Model is evaluated on. For true positive, there is one and only one ground truth annotation, which matches the only prediction in [predictions][google.cloud.aiplatform.v1.EvaluatedAnnotation.predictions]. For false positive, there are zero or more ground truth annotations that are similar to the only prediction in [predictions][google.cloud.aiplatform.v1.EvaluatedAnnotation.predictions], but not enough for a match. For false negative, there is one and only one ground truth annotation, which doesn't match any predictions created by the model. The schema of the ground truth is stored in [ModelEvaluation.annotation_schema_uri][google.cloud.aiplatform.v1.ModelEvaluation.annotation_schema_uri]
Output only. The data item payload that the Model predicted this EvaluatedAnnotation on.
Output only. ID of the EvaluatedDataItemView under the same ancestor ModelEvaluation. The EvaluatedDataItemView consists of all ground truths and predictions on [data_item_payload][google.cloud.aiplatform.v1.EvaluatedAnnotation.data_item_payload].
Explanations of [predictions][google.cloud.aiplatform.v1.EvaluatedAnnotation.predictions]. Each element of the explanations indicates the explanation for one explanation Method. The attributions list in the [EvaluatedAnnotationExplanation.explanation][google.cloud.aiplatform.v1.EvaluatedAnnotationExplanation.explanation] object corresponds to the [predictions][google.cloud.aiplatform.v1.EvaluatedAnnotation.predictions] list. For example, the second element in the attributions list explains the second element in the predictions list.
Annotations of model error analysis results.
Describes the type of the EvaluatedAnnotation. The type is determined
Used in:
Invalid value.
The EvaluatedAnnotation is a true positive. It has a prediction created by the Model and a ground truth Annotation which the prediction matches.
The EvaluatedAnnotation is false positive. It has a prediction created by the Model which does not match any ground truth annotation.
The EvaluatedAnnotation is false negative. It has a ground truth annotation which is not matched by any of the model created predictions.
Explanation result of the prediction produced by the Model.
Used in:
Explanation type. For AutoML Image Classification models, possible values are: * `image-integrated-gradients` * `image-xrai`
Explanation attribution response details.
An edge describing the relationship between an Artifact and an Execution in a lineage graph.
Used in:
,Required. The relative resource name of the Artifact in the Event.
Output only. The relative resource name of the Execution in the Event.
Output only. Time the Event occurred.
Required. The type of the Event.
The labels with user-defined metadata to annotate Events. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Event (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Describes whether an Event's Artifact is the Execution's input or output.
Used in:
Unspecified whether input or output of the Execution.
An input of the Execution.
An output of the Execution.
Input for exact match metric.
Used in:
Required. Spec for exact match metric.
Required. Repeated exact match instances.
Spec for exact match instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Exact match metric value for an instance.
Used in:
Output only. Exact match score.
Results for exact match metric.
Used in:
Output only. Exact match metric values.
Spec for exact match metric - returns 1 if prediction and reference exactly matches, otherwise 0.
Used in:
(message has no fields)
Example-based explainability that returns the nearest neighbors from the provided dataset.
Used in:
,The Cloud Storage input instances.
The full configuration for the generated index, the semantics are the same as [metadata][google.cloud.aiplatform.v1.Index.metadata] and should match [NearestNeighborSearchConfig](https://cloud.google.com/vertex-ai/docs/explainable-ai/configuring-explanations-example-based#nearest-neighbor-search-config).
Simplified preset configuration, which automatically sets configuration values based on the desired query speed-precision trade-off and modality.
The number of neighbors to return when querying for examples.
The Cloud Storage input instances.
Used in:
The format in which instances are given, if not specified, assume it's JSONL format. Currently only JSONL format is supported.
The Cloud Storage location for the input instances.
The format of the input example instances.
Used in:
Format unspecified, used when unset.
Examples are stored in JSONL files.
Overrides for example-based explanations.
Used in:
The number of neighbors to return.
The number of neighbors to return that have the same crowding tag.
Restrict the resulting nearest neighbors to respect these constraints.
If true, return the embeddings instead of neighbors.
The format of the data being provided with each call.
Data format enum.
Used in:
Unspecified format. Must not be used.
Provided data is a set of model inputs.
Provided data is a set of embeddings.
Restrictions namespace for example-based explanations overrides.
Used in:
The namespace name.
The list of allowed tags.
The list of deny tags.
Code generated by the model that is meant to be executed, and the result returned to the model. Generated when using the [FunctionDeclaration] tool and [FunctionCallingConfig] mode is set to [Mode.CODE].
Used in:
Required. Programming language of the `code`.
Required. The code to be executed.
Supported programming languages for the generated code.
Used in:
Unspecified language. This value should not be used.
Python >= 3.10, with numpy and simpy available.
Instance of a general execution.
Used as response type in: MetadataService.CreateExecution, MetadataService.GetExecution, MetadataService.UpdateExecution
Used as field type in:
, , , ,Output only. The resource name of the Execution.
User provided display name of the Execution. May be up to 128 Unicode characters.
The state of this Execution. This is a property of the Execution, and does not imply or capture any ongoing process. This property is managed by clients (such as Vertex AI Pipelines) and the system does not prescribe or check the validity of state transitions.
An eTag used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Executions. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Execution (System labels are excluded).
Output only. Timestamp when this Execution was created.
Output only. Timestamp when this Execution was last updated.
The title of the schema describing the metadata. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
The version of the schema in `schema_title` to use. Schema title and version is expected to be registered in earlier Create Schema calls. And both are used together as unique identifiers to identify schemas within the local metadata store.
Properties of the Execution. Top level metadata keys' heading and trailing spaces will be trimmed. The size of this field should not exceed 200KB.
Description of the Execution
Describes the state of the Execution.
Used in:
Unspecified Execution state
The Execution is new
The Execution is running
The Execution has finished running
The Execution has failed
The Execution completed through Cache hit.
The Execution was cancelled.
Explanation of a prediction (provided in [PredictResponse.predictions][google.cloud.aiplatform.v1.PredictResponse.predictions]) produced by the Model on a given [instance][google.cloud.aiplatform.v1.ExplainRequest.instances].
Used in:
,Output only. Feature attributions grouped by predicted outputs. For Models that predict only one output, such as regression Models that predict only one score, there is only one attibution that explains the predicted output. For Models that predict multiple outputs, such as multiclass Models that predict multiple classes, each element explains one specific item. [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] can be used to identify which output this attribution is explaining. By default, we provide Shapley values for the predicted class. However, you can configure the explanation request to generate Shapley values for any other classes too. For example, if a model predicts a probability of `0.4` for approving a loan application, the model's decision is to reject the application since `p(reject) = 0.6 > p(approve) = 0.4`, and the default Shapley values would be computed for rejection decision and not approval, even though the latter might be the positive class. If users set [ExplanationParameters.top_k][google.cloud.aiplatform.v1.ExplanationParameters.top_k], the attributions are sorted by [instance_output_value][google.cloud.aiplatform.v1.Attribution.instance_output_value] in descending order. If [ExplanationParameters.output_indices][google.cloud.aiplatform.v1.ExplanationParameters.output_indices] is specified, the attributions are stored by [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] in the same order as they appear in the output_indices.
Output only. List of the nearest neighbors for example-based explanations. For models deployed with the examples explanations feature enabled, the attributions field is empty and instead the neighbors field is populated.
Metadata describing the Model's input and output for explanation.
Used in:
Required. Map from feature names to feature input metadata. Keys are the name of the features. Values are the specification of the feature. An empty InputMetadata is valid. It describes a text feature which has the name specified as the key in [ExplanationMetadata.inputs][google.cloud.aiplatform.v1.ExplanationMetadata.inputs]. The baseline of the empty feature is chosen by Vertex AI. For Vertex AI-provided Tensorflow images, the key can be any friendly name of the feature. Once specified, [featureAttributions][google.cloud.aiplatform.v1.Attribution.feature_attributions] are keyed by this key (if not grouped with another feature). For custom images, the key must match with the key in [instance][google.cloud.aiplatform.v1.ExplainRequest.instances].
Required. Map from output names to output metadata. For Vertex AI-provided Tensorflow images, keys can be any user defined string that consists of any UTF-8 characters. For custom images, keys are the name of the output field in the prediction to be explained. Currently only one key is allowed.
Points to a YAML file stored on Google Cloud Storage describing the format of the [feature attributions][google.cloud.aiplatform.v1.Attribution.feature_attributions]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). AutoML tabular Models always have this field populated by Vertex AI. Note: The URI given on output may be different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Name of the source to generate embeddings for example based explanations.
Metadata of the input of a feature. Fields other than [InputMetadata.input_baselines][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.input_baselines] are applicable only for Models that are using Vertex AI-provided images for Tensorflow.
Used in:
Baseline inputs for this feature. If no baseline is specified, Vertex AI chooses the baseline for this feature. If multiple baselines are specified, Vertex AI returns the average attributions across them in [Attribution.feature_attributions][google.cloud.aiplatform.v1.Attribution.feature_attributions]. For Vertex AI-provided Tensorflow images (both 1.x and 2.x), the shape of each baseline must match the shape of the input tensor. If a scalar is provided, we broadcast to the same shape as the input tensor. For custom images, the element of the baselines must be in the same format as the feature's input in the [instance][google.cloud.aiplatform.v1.ExplainRequest.instances][]. The schema of any single instance may be specified via Endpoint's DeployedModels' [Model's][google.cloud.aiplatform.v1.DeployedModel.model] [PredictSchemata's][google.cloud.aiplatform.v1.Model.predict_schemata] [instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri].
Name of the input tensor for this feature. Required and is only applicable to Vertex AI-provided images for Tensorflow.
Defines how the feature is encoded into the input tensor. Defaults to IDENTITY.
Modality of the feature. Valid values are: numeric, image. Defaults to numeric.
The domain details of the input feature value. Like min/max, original mean or standard deviation if normalized.
Specifies the index of the values of the input tensor. Required when the input tensor is a sparse representation. Refer to Tensorflow documentation for more details: https://www.tensorflow.org/api_docs/python/tf/sparse/SparseTensor.
Specifies the shape of the values of the input if the input is a sparse representation. Refer to Tensorflow documentation for more details: https://www.tensorflow.org/api_docs/python/tf/sparse/SparseTensor.
A list of feature names for each index in the input tensor. Required when the input [InputMetadata.encoding][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.encoding] is BAG_OF_FEATURES, BAG_OF_FEATURES_SPARSE, INDICATOR.
Encoded tensor is a transformation of the input tensor. Must be provided if choosing [Integrated Gradients attribution][google.cloud.aiplatform.v1.ExplanationParameters.integrated_gradients_attribution] or [XRAI attribution][google.cloud.aiplatform.v1.ExplanationParameters.xrai_attribution] and the input tensor is not differentiable. An encoded tensor is generated if the input tensor is encoded by a lookup table.
A list of baselines for the encoded tensor. The shape of each baseline should match the shape of the encoded tensor. If a scalar is provided, Vertex AI broadcasts to the same shape as the encoded tensor.
Visualization configurations for image explanation.
Name of the group that the input belongs to. Features with the same group name will be treated as one feature when computing attributions. Features grouped together can have different shapes in value. If provided, there will be one single attribution generated in [Attribution.feature_attributions][google.cloud.aiplatform.v1.Attribution.feature_attributions], keyed by the group name.
Defines how a feature is encoded. Defaults to IDENTITY.
Used in:
Default value. This is the same as IDENTITY.
The tensor represents one feature.
The tensor represents a bag of features where each index maps to a feature. [InputMetadata.index_feature_mapping][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.index_feature_mapping] must be provided for this encoding. For example: ``` input = [27, 6.0, 150] index_feature_mapping = ["age", "height", "weight"] ```
The tensor represents a bag of features where each index maps to a feature. Zero values in the tensor indicates feature being non-existent. [InputMetadata.index_feature_mapping][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.index_feature_mapping] must be provided for this encoding. For example: ``` input = [2, 0, 5, 0, 1] index_feature_mapping = ["a", "b", "c", "d", "e"] ```
The tensor is a list of binaries representing whether a feature exists or not (1 indicates existence). [InputMetadata.index_feature_mapping][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.index_feature_mapping] must be provided for this encoding. For example: ``` input = [1, 0, 1, 0, 1] index_feature_mapping = ["a", "b", "c", "d", "e"] ```
The tensor is encoded into a 1-dimensional array represented by an encoded tensor. [InputMetadata.encoded_tensor_name][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.encoded_tensor_name] must be provided for this encoding. For example: ``` input = ["This", "is", "a", "test", "."] encoded = [0.1, 0.2, 0.3, 0.4, 0.5] ```
Select this encoding when the input tensor is encoded into a 2-dimensional array represented by an encoded tensor. [InputMetadata.encoded_tensor_name][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata.encoded_tensor_name] must be provided for this encoding. The first dimension of the encoded tensor's shape is the same as the input tensor's shape. For example: ``` input = ["This", "is", "a", "test", "."] encoded = [[0.1, 0.2, 0.3, 0.4, 0.5], [0.2, 0.1, 0.4, 0.3, 0.5], [0.5, 0.1, 0.3, 0.5, 0.4], [0.5, 0.3, 0.1, 0.2, 0.4], [0.4, 0.3, 0.2, 0.5, 0.1]] ```
Domain details of the input feature value. Provides numeric information about the feature, such as its range (min, max). If the feature has been pre-processed, for example with z-scoring, then it provides information about how to recover the original feature. For example, if the input feature is an image and it has been pre-processed to obtain 0-mean and stddev = 1 values, then original_mean, and original_stddev refer to the mean and stddev of the original feature (e.g. image tensor) from which input feature (with mean = 0 and stddev = 1) was obtained.
Used in:
The minimum permissible value for this feature.
The maximum permissible value for this feature.
If this input feature has been normalized to a mean value of 0, the original_mean specifies the mean value of the domain prior to normalization.
If this input feature has been normalized to a standard deviation of 1.0, the original_stddev specifies the standard deviation of the domain prior to normalization.
Visualization configurations for image explanation.
Used in:
Type of the image visualization. Only applicable to [Integrated Gradients attribution][google.cloud.aiplatform.v1.ExplanationParameters.integrated_gradients_attribution]. OUTLINES shows regions of attribution, while PIXELS shows per-pixel attribution. Defaults to OUTLINES.
Whether to only highlight pixels with positive contributions, negative or both. Defaults to POSITIVE.
The color scheme used for the highlighted areas. Defaults to PINK_GREEN for [Integrated Gradients attribution][google.cloud.aiplatform.v1.ExplanationParameters.integrated_gradients_attribution], which shows positive attributions in green and negative in pink. Defaults to VIRIDIS for [XRAI attribution][google.cloud.aiplatform.v1.ExplanationParameters.xrai_attribution], which highlights the most influential regions in yellow and the least influential in blue.
Excludes attributions above the specified percentile from the highlighted areas. Using the clip_percent_upperbound and clip_percent_lowerbound together can be useful for filtering out noise and making it easier to see areas of strong attribution. Defaults to 99.9.
Excludes attributions below the specified percentile, from the highlighted areas. Defaults to 62.
How the original image is displayed in the visualization. Adjusting the overlay can help increase visual clarity if the original image makes it difficult to view the visualization. Defaults to NONE.
The color scheme used for highlighting areas.
Used in:
Should not be used.
Positive: green. Negative: pink.
Viridis color map: A perceptually uniform color mapping which is easier to see by those with colorblindness and progresses from yellow to green to blue. Positive: yellow. Negative: blue.
Positive: red. Negative: red.
Positive: green. Negative: green.
Positive: green. Negative: red.
PiYG palette.
How the original image is displayed in the visualization.
Used in:
Default value. This is the same as NONE.
No overlay.
The attributions are shown on top of the original image.
The attributions are shown on top of grayscaled version of the original image.
The attributions are used as a mask to reveal predictive parts of the image and hide the un-predictive parts.
Whether to only highlight pixels with positive contributions, negative or both. Defaults to POSITIVE.
Used in:
Default value. This is the same as POSITIVE.
Highlights the pixels/outlines that were most influential to the model's prediction.
Setting polarity to negative highlights areas that does not lead to the models's current prediction.
Shows both positive and negative attributions.
Type of the image visualization. Only applicable to [Integrated Gradients attribution][google.cloud.aiplatform.v1.ExplanationParameters.integrated_gradients_attribution].
Used in:
Should not be used.
Shows which pixel contributed to the image prediction.
Shows which region contributed to the image prediction by outlining the region.
Metadata of the prediction output to be explained.
Used in:
Defines how to map [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] to [Attribution.output_display_name][google.cloud.aiplatform.v1.Attribution.output_display_name]. If neither of the fields are specified, [Attribution.output_display_name][google.cloud.aiplatform.v1.Attribution.output_display_name] will not be populated.
Static mapping between the index and display name. Use this if the outputs are a deterministic n-dimensional array, e.g. a list of scores of all the classes in a pre-defined order for a multi-classification Model. It's not feasible if the outputs are non-deterministic, e.g. the Model produces top-k classes or sort the outputs by their values. The shape of the value must be an n-dimensional array of strings. The number of dimensions must match that of the outputs to be explained. The [Attribution.output_display_name][google.cloud.aiplatform.v1.Attribution.output_display_name] is populated by locating in the mapping with [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index].
Specify a field name in the prediction to look for the display name. Use this if the prediction contains the display names for the outputs. The display names in the prediction must have the same shape of the outputs, so that it can be located by [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] for a specific output.
Name of the output tensor. Required and is only applicable to Vertex AI provided images for Tensorflow.
The [ExplanationMetadata][google.cloud.aiplatform.v1.ExplanationMetadata] entries that can be overridden at [online explanation][google.cloud.aiplatform.v1.PredictionService.Explain] time.
Used in:
Required. Overrides the [input metadata][google.cloud.aiplatform.v1.ExplanationMetadata.inputs] of the features. The key is the name of the feature to be overridden. The keys specified here must exist in the input metadata to be overridden. If a feature is not specified here, the corresponding feature's input metadata is not overridden.
The [input metadata][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata] entries to be overridden.
Used in:
Baseline inputs for this feature. This overrides the `input_baseline` field of the [ExplanationMetadata.InputMetadata][google.cloud.aiplatform.v1.ExplanationMetadata.InputMetadata] object of the corresponding feature's input metadata. If it's not specified, the original baselines are not overridden.
Parameters to configure explaining for Model's predictions.
Used in:
,An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features. Refer to this paper for model details: https://arxiv.org/abs/1306.4265.
An attribution method that computes Aumann-Shapley values taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
An attribution method that redistributes Integrated Gradients attribution to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 XRAI currently performs better on natural images, like a picture of a house or an animal. If the images are taken in artificial environments, like a lab or manufacturing line, or from diagnostic equipment, like x-rays or quality-control cameras, use Integrated Gradients instead.
Example-based explanations that returns the nearest neighbors from the provided dataset.
If populated, returns attributions for top K indices of outputs (defaults to 1). Only applies to Models that predicts more than one outputs (e,g, multi-class Models). When set to -1, returns explanations for all outputs.
If populated, only returns attributions that have [output_index][google.cloud.aiplatform.v1.Attribution.output_index] contained in output_indices. It must be an ndarray of integers, with the same shape of the output it's explaining. If not populated, returns attributions for [top_k][google.cloud.aiplatform.v1.ExplanationParameters.top_k] indices of outputs. If neither top_k nor output_indices is populated, returns the argmax index of the outputs. Only applicable to Models that predict multiple outputs (e,g, multi-class Models that predict multiple classes).
Specification of Model explanation.
Used in:
, , ,Required. Parameters that configure explaining of the Model's predictions.
Optional. Metadata describing the Model's input and output for explanation.
The [ExplanationSpec][google.cloud.aiplatform.v1.ExplanationSpec] entries that can be overridden at [online explanation][google.cloud.aiplatform.v1.PredictionService.Explain] time.
Used in:
The parameters to be overridden. Note that the attribution method cannot be changed. If not specified, no parameter is overridden.
The metadata to be overridden. If not specified, no metadata is overridden.
The example-based explanations parameter overrides.
Describes what part of the Dataset is to be exported, the destination of the export and how to export.
Used in:
The destination of the output.
The Google Cloud Storage location where the output is to be written to. In the given directory a new directory will be created with name: `export-data-<dataset-display-name>-<timestamp-of-export-call>` where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. All export output will be written into that directory. Inside that directory, annotations with the same schema will be grouped into sub directories which are named with the corresponding annotations' schema title. Inside these sub directories, a schema.yaml will be created to describe the output format.
The instructions how the export data should be split between the training, validation and test sets.
Split based on fractions defining the size of each set.
Split based on the provided filters for each set.
An expression for filtering what part of the Dataset is to be exported. Only Annotations that match this filter will be exported. The filter syntax is the same as in [ListAnnotations][google.cloud.aiplatform.v1.DatasetService.ListAnnotations].
The ID of a SavedQuery (annotation set) under the Dataset specified by [ExportDataRequest.name][google.cloud.aiplatform.v1.ExportDataRequest.name] used for filtering Annotations for training. Only used for custom training data export use cases. Only applicable to Datasets that have SavedQueries. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1.ExportDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1.ExportDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1.ExportDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1.ExportDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1.ExportDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
The Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/, note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri] of the Dataset specified by [ExportDataRequest.name][google.cloud.aiplatform.v1.ExportDataRequest.name]. Only used for custom training data export use cases. Only applicable to Datasets that have DataItems and Annotations. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1.ExportDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1.ExportDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1.ExportDataConfig.annotation_schema_uri].
Indicates the usage of the exported files.
ExportUse indicates the usage of the exported files. It restricts file destination, format, annotations to be exported, whether to allow unannotated data to be exported and whether to clone files to temp Cloud Storage bucket.
Used in:
Regular user export.
Export for custom code training.
Runtime operation information for [DatasetService.ExportData][google.cloud.aiplatform.v1.DatasetService.ExportData].
The common part of the operation metadata.
A Google Cloud Storage directory which path ends with '/'. The exported data is stored in the directory.
Response message for [DatasetService.ExportData][google.cloud.aiplatform.v1.DatasetService.ExportData].
All of the files that are exported in this export operation. For custom code training export, only three (training, validation and test) Cloud Storage paths in wildcard format are populated (for example, gs://.../training-*).
Only present for custom code training export use case. Records data stats, i.e., train/validation/test item/annotation counts calculated during the export operation.
Details of operations that exports Features values.
Operation metadata for Featurestore export Feature values.
Describes exporting all historical Feature values of all entities of the EntityType between [start_time, end_time].
Used in:
Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision.
Exports Feature values as of this timestamp. If not set, retrieve values as of now. Timestamp, if present, must not have higher than millisecond precision.
Describes exporting the latest Feature values of all entities of the EntityType between [start_time, snapshot_time].
Used in:
Exports Feature values as of this timestamp. If not set, retrieve values as of now. Timestamp, if present, must not have higher than millisecond precision.
Excludes Feature values with feature generation timestamp before this timestamp. If not set, retrieve oldest values kept in Feature Store. Timestamp, if present, must not have higher than millisecond precision.
Response message for [FeaturestoreService.ExportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ExportFeatureValues].
(message has no fields)
Assigns input data to training, validation, and test sets based on the given filters, data pieces not matched by any filter are ignored. Currently only supported for Datasets containing DataItems. If any of the filters in this message are to match nothing, then they can be set as '-' (the minus sign). Supported only for unstructured Datasets.
Used in:
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to train the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to validate the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to test the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
Assigns the input data to training, validation, and test sets as per the given fractions. Any of `training_fraction`, `validation_fraction` and `test_fraction` may optionally be provided, they must sum to up to 1. If the provided ones sum to less than 1, the remainder is assigned to sets as decided by Vertex AI. If none of the fractions are set, by default roughly 80% of data is used for training, 10% for validation, and 10% for test.
Used in:
The fraction of the input data that is to be used to train the Model.
The fraction of the input data that is to be used to validate the Model.
The fraction of the input data that is to be used to evaluate the Model.
Details of [ModelService.ExportModel][google.cloud.aiplatform.v1.ModelService.ExportModel] operation.
The common part of the operation metadata.
Output only. Information further describing the output of this Model export.
Further describes the output of the ExportModel. Supplements [ExportModelRequest.OutputConfig][google.cloud.aiplatform.v1.ExportModelRequest.OutputConfig].
Used in:
Output only. If the Model artifact is being exported to Google Cloud Storage this is the full path of the directory created, into which the Model files are being written to.
Output only. If the Model image is being exported to Google Container Registry or Artifact Registry this is the full path of the image created.
Output configuration for the Model export.
Used in:
The ID of the format in which the Model must be exported. Each Model lists the [export formats it supports][google.cloud.aiplatform.v1.Model.supported_export_formats]. If no value is provided here, then the first from the list of the Model's supported formats is used by default.
The Cloud Storage location where the Model artifact is to be written to. Under the directory given as the destination a new one with name "`model-export-<model-display-name>-<timestamp-of-export-call>`", where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format, will be created. Inside, the Model and any of its supporting files will be written. This field should only be set when the `exportableContent` field of the [Model.supported_export_formats] object contains `ARTIFACT`.
The Google Container Registry or Artifact Registry uri where the Model container image will be copied to. This field should only be set when the `exportableContent` field of the [Model.supported_export_formats] object contains `IMAGE`.
Response message of [ModelService.ExportModel][google.cloud.aiplatform.v1.ModelService.ExportModel] operation.
(message has no fields)
The fact used in grounding.
Used in:
,Query that is used to retrieve this fact.
If present, it refers to the title of this fact.
If present, this uri links to the source of the fact.
If present, the summary/snippet of the fact.
If present, the distance between the query vector and this fact vector.
If present, according to the underlying Vector DB and the selected metric type, the score can be either the distance or the similarity between the query and the fact and its range depends on the metric type. For example, if the metric type is COSINE_DISTANCE, it represents the distance between the query and the fact. The larger the distance, the less relevant the fact is to the query. The range is [0, 2], while 0 means the most relevant and 2 means the least relevant.
If present, chunk properties.
Configuration for faster model deployment.
Used in:
If true, enable fast tryout feature for this deployed model.
Feature Metadata information. For example, color is a feature that describes an apple.
Used as response type in: FeatureRegistryService.GetFeature, FeaturestoreService.GetFeature, FeaturestoreService.UpdateFeature
Used as field type in:
, , , ,Immutable. Name of the Feature. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}/features/{feature}` `projects/{project}/locations/{location}/featureGroups/{feature_group}/features/{feature}` The last part feature is assigned by the client. The feature can be up to 64 characters long and can consist only of ASCII Latin letters A-Z and a-z, underscore(_), and ASCII digits 0-9 starting with a letter. The value will be unique given an entity type.
Description of the Feature.
Immutable. Only applicable for Vertex AI Feature Store (Legacy). Type of Feature value.
Output only. Only applicable for Vertex AI Feature Store (Legacy). Timestamp when this EntityType was created.
Output only. Only applicable for Vertex AI Feature Store (Legacy). Timestamp when this EntityType was most recently updated.
Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type ([Feature.ValueType][google.cloud.aiplatform.v1.Feature.ValueType]) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType.
Output only. Only applicable for Vertex AI Feature Store (Legacy). The list of historical stats and anomalies with specified objectives.
Only applicable for Vertex AI Feature Store. The name of the BigQuery Table/View column hosting data for this version. If no value is provided, will use feature_id.
Entity responsible for maintaining this feature. Can be comma separated list of email addresses or URIs.
A list of historical [SnapshotAnalysis][google.cloud.aiplatform.v1.FeaturestoreMonitoringConfig.SnapshotAnalysis] or [ImportFeaturesAnalysis][google.cloud.aiplatform.v1.FeaturestoreMonitoringConfig.ImportFeaturesAnalysis] stats requested by user, sorted by [FeatureStatsAnomaly.start_time][google.cloud.aiplatform.v1.FeatureStatsAnomaly.start_time] descending.
Used in:
Output only. The objective for each stats.
Output only. The stats and anomalies generated at specific timestamp.
If the objective in the request is both Import Feature Analysis and Snapshot Analysis, this objective could be one of them. Otherwise, this objective should be the same as the objective in the request.
Used in:
If it's OBJECTIVE_UNSPECIFIED, monitoring_stats will be empty.
Stats are generated by Import Feature Analysis.
Stats are generated by Snapshot Analysis.
Only applicable for Vertex AI Legacy Feature Store. An enum representing the value type of a feature.
Used in:
The value type is unspecified.
Used for Feature that is a boolean.
Used for Feature that is a list of boolean.
Used for Feature that is double.
Used for Feature that is a list of double.
Used for Feature that is INT64.
Used for Feature that is a list of INT64.
Used for Feature that is string.
Used for Feature that is a list of String.
Used for Feature that is bytes.
Used for Feature that is struct.
Vertex AI Feature Group.
Used as response type in: FeatureRegistryService.GetFeatureGroup
Used as field type in:
, ,Indicates that features for this group come from BigQuery Table/View. By default treats the source as a sparse time series source. The BigQuery source table or view must have at least one entity ID column and a column named `feature_timestamp`.
Identifier. Name of the FeatureGroup. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}`
Output only. Timestamp when this FeatureGroup was created.
Output only. Timestamp when this FeatureGroup was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. The labels with user-defined metadata to organize your FeatureGroup. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureGroup(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Optional. Description of the FeatureGroup.
Input source type for BigQuery Tables and Views.
Used in:
Required. Immutable. The BigQuery source URI that points to either a BigQuery Table or View.
Optional. Columns to construct entity_id / row keys. If not provided defaults to `entity_id`.
Optional. Set if the data source is not a time-series.
Optional. If the source is a time-series source, this can be set to control how downstream sources (ex: [FeatureView][google.cloud.aiplatform.v1.FeatureView] ) will treat time-series sources. If not set, will treat the source as a time-series source with `feature_timestamp` as timestamp column and no scan boundary.
Optional. If set, all feature values will be fetched from a single row per unique entityId including nulls. If not set, will collapse all rows for each unique entityId into a singe row with any non-null values if present, if no non-null values are present will sync null. ex: If source has schema `(entity_id, feature_timestamp, f0, f1)` and the following rows: `(e1, 2020-01-01T10:00:00.123Z, 10, 15)` `(e1, 2020-02-01T10:00:00.123Z, 20, null)` If dense is set, `(e1, 20, null)` is synced to online stores. If dense is not set, `(e1, 20, 15)` is synced to online stores.
Used in:
Optional. Column hosting timestamp values for a time-series source. Will be used to determine the latest `feature_values` for each entity. Optional. If not provided, column named `feature_timestamp` of type `TIMESTAMP` will be used.
Noise sigma by features. Noise sigma represents the standard deviation of the gaussian kernel that will be used to add noise to interpolated inputs prior to computing gradients.
Used in:
Noise sigma per feature. No noise is added to features that are not set.
Noise sigma for a single feature.
Used in:
The name of the input feature for which noise sigma is provided. The features are defined in [explanation metadata inputs][google.cloud.aiplatform.v1.ExplanationMetadata.inputs].
This represents the standard deviation of the Gaussian kernel that will be used to add noise to the feature prior to computing gradients. Similar to [noise_sigma][google.cloud.aiplatform.v1.SmoothGradConfig.noise_sigma] but represents the noise added to the current feature. Defaults to 0.1.
Vertex AI Feature Online Store provides a centralized repository for serving ML features and embedding indexes at low latency. The Feature Online Store is a top-level container.
Used as response type in: FeatureOnlineStoreAdminService.GetFeatureOnlineStore
Used as field type in:
, ,Contains settings for the Cloud Bigtable instance that will be created to serve featureValues for all FeatureViews under this FeatureOnlineStore.
Contains settings for the Optimized store that will be created to serve featureValues for all FeatureViews under this FeatureOnlineStore. When choose Optimized storage type, need to set [PrivateServiceConnectConfig.enable_private_service_connect][google.cloud.aiplatform.v1.PrivateServiceConnectConfig.enable_private_service_connect] to use private endpoint. Otherwise will use public endpoint by default.
Identifier. Name of the FeatureOnlineStore. Format: `projects/{project}/locations/{location}/featureOnlineStores/{featureOnlineStore}`
Output only. Timestamp when this FeatureOnlineStore was created.
Output only. Timestamp when this FeatureOnlineStore was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. The labels with user-defined metadata to organize your FeatureOnlineStore. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureOnlineStore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Output only. State of the featureOnlineStore.
Optional. The dedicated serving endpoint for this FeatureOnlineStore, which is different from common Vertex service endpoint.
Optional. Customer-managed encryption key spec for data storage. If set, online store will be secured by this key.
Output only. Reserved for future use.
Output only. Reserved for future use.
Used in:
Required. Autoscaling config applied to Bigtable Instance.
Used in:
Required. The minimum number of nodes to scale down to. Must be greater than or equal to 1.
Required. The maximum number of nodes to scale up to. Must be greater than or equal to min_node_count, and less than or equal to 10 times of 'min_node_count'.
Optional. A percentage of the cluster's CPU capacity. Can be from 10% to 80%. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set will default to 50%.
The dedicated serving endpoint for this FeatureOnlineStore. Only need to set when you choose Optimized storage type. Public endpoint is provisioned by default.
Used in:
Output only. This field will be populated with the domain name to use for this FeatureOnlineStore
Optional. Private service connect config. The private service connection is available only for Optimized storage type, not for embedding management now. If [PrivateServiceConnectConfig.enable_private_service_connect][google.cloud.aiplatform.v1.PrivateServiceConnectConfig.enable_private_service_connect] set to true, customers will use private service connection to send request. Otherwise, the connection will set to public endpoint.
Output only. The name of the service attachment resource. Populated if private service connect is enabled and after FeatureViewSync is created.
Optimized storage type
Used in:
(message has no fields)
Possible states a featureOnlineStore can have.
Used in:
Default value. This value is unused.
State when the featureOnlineStore configuration is not being updated and the fields reflect the current configuration of the featureOnlineStore. The featureOnlineStore is usable in this state.
The state of the featureOnlineStore configuration when it is being updated. During an update, the fields reflect either the original configuration or the updated configuration of the featureOnlineStore. The featureOnlineStore is still usable in this state.
Selector for Features of an EntityType.
Used in:
, , , ,Required. Matches Features based on ID.
Stats and Anomaly generated at specific timestamp for specific Feature. The start_time and end_time are used to define the time range of the dataset that current stats belongs to, e.g. prediction traffic is bucketed into prediction datasets by time window. If the Dataset is not defined by time window, start_time = end_time. Timestamp of the stats and anomalies always refers to end_time. Raw stats and anomalies are stored in stats_uri or anomaly_uri in the tensorflow defined protos. Field data_stats contains almost identical information with the raw stats in Vertex AI defined proto, for UI to display.
Used in:
,Feature importance score, only populated when cross-feature monitoring is enabled. For now only used to represent feature attribution score within range [0, 1] for [ModelDeploymentMonitoringObjectiveType.FEATURE_ATTRIBUTION_SKEW][google.cloud.aiplatform.v1.ModelDeploymentMonitoringObjectiveType.FEATURE_ATTRIBUTION_SKEW] and [ModelDeploymentMonitoringObjectiveType.FEATURE_ATTRIBUTION_DRIFT][google.cloud.aiplatform.v1.ModelDeploymentMonitoringObjectiveType.FEATURE_ATTRIBUTION_DRIFT].
Path of the stats file for current feature values in Cloud Storage bucket. Format: gs://<bucket_name>/<object_name>/stats. Example: gs://monitoring_bucket/feature_name/stats. Stats are stored as binary format with Protobuf message [tensorflow.metadata.v0.FeatureNameStatistics](https://github.com/tensorflow/metadata/blob/master/tensorflow_metadata/proto/v0/statistics.proto).
Path of the anomaly file for current feature values in Cloud Storage bucket. Format: gs://<bucket_name>/<object_name>/anomalies. Example: gs://monitoring_bucket/feature_name/anomalies. Stats are stored as binary format with Protobuf message Anoamlies are stored as binary format with Protobuf message [tensorflow.metadata.v0.AnomalyInfo] (https://github.com/tensorflow/metadata/blob/master/tensorflow_metadata/proto/v0/anomalies.proto).
Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence.
This is the threshold used when detecting anomalies. The threshold can be changed by user, so this one might be different from [ThresholdConfig.value][google.cloud.aiplatform.v1.ThresholdConfig.value].
The start timestamp of window where stats were generated. For objectives where time window doesn't make sense (e.g. Featurestore Snapshot Monitoring), start_time is only used to indicate the monitoring intervals, so it always equals to (end_time - monitoring_interval).
The end timestamp of window where stats were generated. For objectives where time window doesn't make sense (e.g. Featurestore Snapshot Monitoring), end_time indicates the timestamp of the data used to generate stats (e.g. timestamp we take snapshots for feature values).
Value for a feature.
Used in:
, , , ,Value for the feature.
Bool type feature value.
Double type feature value.
Int64 feature value.
String feature value.
A list of bool type feature value.
A list of double type feature value.
A list of int64 type feature value.
A list of string type feature value.
Bytes feature value.
A struct type feature value.
Metadata of feature value.
Metadata of feature value.
Used in:
Feature generation timestamp. Typically, it is provided by user at feature ingestion time. If not, feature store will use the system timestamp when the data is ingested into feature store. For streaming ingestion, the time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.
A destination location for Feature values and format.
Used in:
,Output in BigQuery format. [BigQueryDestination.output_uri][google.cloud.aiplatform.v1.BigQueryDestination.output_uri] in [FeatureValueDestination.bigquery_destination][google.cloud.aiplatform.v1.FeatureValueDestination.bigquery_destination] must refer to a table.
Output in TFRecord format. Below are the mapping from Feature value type in Featurestore to Feature value type in TFRecord: Value type in Featurestore | Value type in TFRecord DOUBLE, DOUBLE_ARRAY | FLOAT_LIST INT64, INT64_ARRAY | INT64_LIST STRING, STRING_ARRAY, BYTES | BYTES_LIST true -> byte_string("true"), false -> byte_string("false") BOOL, BOOL_ARRAY (true, false) | BYTES_LIST
Output in CSV format. Array Feature value types are not allowed in CSV format.
Container for list of values.
Used in:
A list of feature values. All of them should be the same data type.
FeatureView is representation of values that the FeatureOnlineStore will serve based on its syncConfig.
Used as response type in: FeatureOnlineStoreAdminService.GetFeatureView
Used as field type in:
, ,Optional. Configures how data is supposed to be extracted from a BigQuery source to be loaded onto the FeatureOnlineStore.
Optional. Configures the features from a Feature Registry source that need to be loaded onto the FeatureOnlineStore.
Optional. The Vertex RAG Source that the FeatureView is linked to.
Identifier. Name of the FeatureView. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}`
Output only. Timestamp when this FeatureView was created.
Output only. Timestamp when this FeatureView was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. The labels with user-defined metadata to organize your FeatureViews. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureOnlineStore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Configures when data is to be synced/updated for this FeatureView. At the end of the sync the latest featureValues for each entityId of this FeatureView are made ready for online serving.
Optional. Configuration for index preparation for vector search. It contains the required configurations to create an index from source data, so that approximate nearest neighbor (a.k.a ANN) algorithms search can be performed during online serving.
Optional. Configuration for FeatureView created under Optimized FeatureOnlineStore.
Optional. Service agent type used during data sync. By default, the Vertex AI Service Agent is used. When using an IAM Policy to isolate this FeatureView within a project, a separate service account should be provisioned by setting this field to `SERVICE_AGENT_TYPE_FEATURE_VIEW`. This will generate a separate service account to access the BigQuery source table.
Output only. A Service Account unique to this FeatureView. The role bigquery.dataViewer should be granted to this service account to allow Vertex AI Feature Store to sync data to the online store.
Output only. Reserved for future use.
Output only. Reserved for future use.
Used in:
Required. The BigQuery view URI that will be materialized on each sync trigger based on FeatureView.SyncConfig.
Required. Columns to construct entity_id / row keys.
A Feature Registry source for features that need to be synced to Online Store.
Used in:
Required. List of features that need to be synced to Online Store.
Optional. The project number of the parent project of the Feature Groups.
Features belonging to a single feature group that will be synced to Online Store.
Used in:
Required. Identifier of the feature group.
Required. Identifiers of features under the feature group.
Configuration for vector indexing.
Used in:
The configuration with regard to the algorithms used for efficient search.
Optional. Configuration options for the tree-AH algorithm (Shallow tree + Asymmetric Hashing). Please refer to this paper for more details: https://arxiv.org/abs/1908.10396
Optional. Configuration options for using brute force search, which simply implements the standard linear search in the database for each query. It is primarily meant for benchmarking and to generate the ground truth for approximate search.
Optional. Column of embedding. This column contains the source data to create index for vector search. embedding_column must be set when using vector search.
Optional. Columns of features that're used to filter vector search results.
Optional. Column of crowding. This column contains crowding attribute which is a constraint on a neighbor list produced by [FeatureOnlineStoreService.SearchNearestEntities][google.cloud.aiplatform.v1.FeatureOnlineStoreService.SearchNearestEntities] to diversify search results. If [NearestNeighborQuery.per_crowding_attribute_neighbor_count][google.cloud.aiplatform.v1.NearestNeighborQuery.per_crowding_attribute_neighbor_count] is set to K in [SearchNearestEntitiesRequest][google.cloud.aiplatform.v1.SearchNearestEntitiesRequest], it's guaranteed that no more than K entities of the same crowding attribute are returned in the response.
Optional. The number of dimensions of the input embedding.
Optional. The distance measure used in nearest neighbor search.
Configuration options for using brute force search.
Used in:
(message has no fields)
The distance measure used in nearest neighbor search.
Used in:
Should not be set.
Euclidean (L_2) Distance.
Cosine Distance. Defined as 1 - cosine similarity. We strongly suggest using DOT_PRODUCT_DISTANCE + UNIT_L2_NORM instead of COSINE distance. Our algorithms have been more optimized for DOT_PRODUCT distance which, when combined with UNIT_L2_NORM, is mathematically equivalent to COSINE distance and results in the same ranking.
Dot Product Distance. Defined as a negative of the dot product.
Configuration options for the tree-AH algorithm.
Used in:
Optional. Number of embeddings on each leaf node. The default value is 1000 if not set.
Configuration for FeatureViews created in Optimized FeatureOnlineStore.
Used in:
Optional. A description of resources that the FeatureView uses, which to large degree are decided by Vertex AI, and optionally allows only a modest additional configuration. If min_replica_count is not set, the default value is 2. If max_replica_count is not set, the default value is 6. The max allowed replica count is 1000.
Service agent type used during data sync.
Used in:
By default, the project-level Vertex AI Service Agent is enabled.
Indicates the project-level Vertex AI Service Agent (https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents) will be used during sync jobs.
Enable a FeatureView service account to be created by Vertex AI and output in the field `service_account_email`. This service account will be used to read from the source BigQuery table during sync.
Configuration for Sync. Only one option is set.
Used in:
Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
Optional. If true, syncs the FeatureView in a continuous manner to Online Store.
A Vertex Rag source for features that need to be synced to Online Store.
Used in:
Required. The BigQuery view/table URI that will be materialized on each manual sync trigger. The table/view is expected to have the following columns and types at least: - `corpus_id` (STRING, NULLABLE/REQUIRED) - `file_id` (STRING, NULLABLE/REQUIRED) - `chunk_id` (STRING, NULLABLE/REQUIRED) - `chunk_data_type` (STRING, NULLABLE/REQUIRED) - `chunk_data` (STRING, NULLABLE/REQUIRED) - `embeddings` (FLOAT, REPEATED) - `file_original_uri` (STRING, NULLABLE/REQUIRED)
Optional. The RAG corpus id corresponding to this FeatureView.
Format of the data in the Feature View.
Used in:
Not set. Will be treated as the KeyValue format.
Return response data in key-value format.
Return response data in proto Struct format.
Lookup key for a feature view.
Used in:
,String key to use for lookup.
The actual Entity ID will be composed from this struct. This should match with the way ID is defined in the FeatureView spec.
ID that is comprised from several parts (columns).
Used in:
Parts to construct Entity ID. Should match with the same ID columns as defined in FeatureView in the same order.
FeatureViewSync is a representation of sync operation which copies data from data source to Feature View in Online Store.
Used as response type in: FeatureOnlineStoreAdminService.GetFeatureViewSync
Used as field type in:
Identifier. Name of the FeatureViewSync. Format: `projects/{project}/locations/{location}/featureOnlineStores/{feature_online_store}/featureViews/{feature_view}/featureViewSyncs/{feature_view_sync}`
Output only. Time when this FeatureViewSync is created. Creation of a FeatureViewSync means that the job is pending / waiting for sufficient resources but may not have started the actual data transfer yet.
Output only. Time when this FeatureViewSync is finished.
Output only. Final status of the FeatureViewSync.
Output only. Summary of the sync job.
Output only. Reserved for future use.
Output only. Reserved for future use.
Summary from the Sync job. For continuous syncs, the summary is updated periodically. For batch syncs, it gets updated on completion of the sync.
Used in:
Output only. Total number of rows synced.
Output only. BigQuery slot milliseconds consumed for the sync job.
Lower bound of the system time watermark for the sync job. This is only set for continuously syncing feature views.
Vertex AI Feature Store provides a centralized repository for organizing, storing, and serving ML features. The Featurestore is a top-level container for your features and their values.
Used as response type in: FeaturestoreService.GetFeaturestore
Used as field type in:
, ,Output only. Name of the Featurestore. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}`
Output only. Timestamp when this Featurestore was created.
Output only. Timestamp when this Featurestore was last updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Optional. The labels with user-defined metadata to organize your Featurestore. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Featurestore(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Optional. Config for online storage resources. The field should not co-exist with the field of `OnlineStoreReplicationConfig`. If both of it and OnlineStoreReplicationConfig are unset, the feature store will not have an online store and cannot be used for online serving.
Output only. State of the featurestore.
Optional. TTL in days for feature values that will be stored in online serving storage. The Feature Store online storage periodically removes obsolete feature values older than `online_storage_ttl_days` since the feature generation time. Note that `online_storage_ttl_days` should be less than or equal to `offline_storage_ttl_days` for each EntityType under a featurestore. If not set, default to 4000 days
Optional. Customer-managed encryption key spec for data storage. If set, both of the online and offline data storage will be secured by this key.
Output only. Reserved for future use.
Output only. Reserved for future use.
OnlineServingConfig specifies the details for provisioning online serving resources.
Used in:
The number of nodes for the online store. The number of nodes doesn't scale automatically, but you can manually update the number of nodes. If set to 0, the featurestore will not have an online store and cannot be used for online serving.
Online serving scaling configuration. Only one of `fixed_node_count` and `scaling` can be set. Setting one will reset the other.
Online serving scaling configuration. If min_node_count and max_node_count are set to the same value, the cluster will be configured with the fixed number of node (no auto-scaling).
Used in:
Required. The minimum number of nodes to scale down to. Must be greater than or equal to 1.
The maximum number of nodes to scale up to. Must be greater than min_node_count, and less than or equal to 10 times of 'min_node_count'.
Optional. The cpu utilization that the Autoscaler should be trying to achieve. This number is on a scale from 0 (no utilization) to 100 (total utilization), and is limited between 10 and 80. When a cluster's CPU utilization exceeds the target that you have set, Bigtable immediately adds nodes to the cluster. When CPU utilization is substantially lower than the target, Bigtable removes nodes. If not set or set to 0, default to 50.
Possible states a featurestore can have.
Used in:
Default value. This value is unused.
State when the featurestore configuration is not being updated and the fields reflect the current configuration of the featurestore. The featurestore is usable in this state.
The state of the featurestore configuration when it is being updated. During an update, the fields reflect either the original configuration or the updated configuration of the featurestore. For example, `online_serving_config.fixed_node_count` can take minutes to update. While the update is in progress, the featurestore is in the UPDATING state, and the value of `fixed_node_count` can be the original value or the updated value, depending on the progress of the operation. Until the update completes, the actual number of nodes can still be the original value of `fixed_node_count`. The featurestore is still usable in this state.
Configuration of how features in Featurestore are monitored.
Used in:
The config for Snapshot Analysis Based Feature Monitoring.
The config for ImportFeatures Analysis Based Feature Monitoring.
Threshold for numerical features of anomaly detection. This is shared by all objectives of Featurestore Monitoring for numerical features (i.e. Features with type ([Feature.ValueType][google.cloud.aiplatform.v1.Feature.ValueType]) DOUBLE or INT64).
Threshold for categorical features of anomaly detection. This is shared by all types of Featurestore Monitoring for categorical features (i.e. Features with type ([Feature.ValueType][google.cloud.aiplatform.v1.Feature.ValueType]) BOOL or STRING).
Configuration of the Featurestore's ImportFeature Analysis Based Monitoring. This type of analysis generates statistics for values of each Feature imported by every [ImportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ImportFeatureValues] operation.
Used in:
Whether to enable / disable / inherite default hebavior for import features analysis.
The baseline used to do anomaly detection for the statistics generated by import features analysis.
Defines the baseline to do anomaly detection for feature values imported by each [ImportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ImportFeatureValues] operation.
Used in:
Should not be used.
Choose the later one statistics generated by either most recent snapshot analysis or previous import features analysis. If non of them exists, skip anomaly detection and only generate a statistics.
Use the statistics generated by the most recent snapshot analysis if exists.
Use the statistics generated by the previous import features analysis if exists.
The state defines whether to enable ImportFeature analysis.
Used in:
Should not be used.
The default behavior of whether to enable the monitoring. EntityType-level config: disabled. Feature-level config: inherited from the configuration of EntityType this Feature belongs to.
Explicitly enables import features analysis. EntityType-level config: by default enables import features analysis for all Features under it. Feature-level config: enables import features analysis regardless of the EntityType-level config.
Explicitly disables import features analysis. EntityType-level config: by default disables import features analysis for all Features under it. Feature-level config: disables import features analysis regardless of the EntityType-level config.
Configuration of the Featurestore's Snapshot Analysis Based Monitoring. This type of analysis generates statistics for each Feature based on a snapshot of the latest feature value of each entities every monitoring_interval.
Used in:
The monitoring schedule for snapshot analysis. For EntityType-level config: unset / disabled = true indicates disabled by default for Features under it; otherwise by default enable snapshot analysis monitoring with monitoring_interval for Features under it. Feature-level config: disabled = true indicates disabled regardless of the EntityType-level config; unset monitoring_interval indicates going with EntityType-level config; otherwise run snapshot analysis monitoring with monitoring_interval regardless of the EntityType-level config. Explicitly Disable the snapshot analysis based monitoring.
Configuration of the snapshot analysis based monitoring pipeline running interval. The value indicates number of days.
Customized export features time window for snapshot analysis. Unit is one day. Default value is 3 weeks. Minimum value is 1 day. Maximum value is 4000 days.
The config for Featurestore Monitoring threshold.
Used in:
Specify a threshold value that can trigger the alert. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. Each feature must have a non-zero threshold if they need to be monitored. Otherwise no alert will be triggered for that feature.
Response message for [FeatureOnlineStoreService.FetchFeatureValues][google.cloud.aiplatform.v1.FeatureOnlineStoreService.FetchFeatureValues]
Used as response type in: FeatureOnlineStoreService.FetchFeatureValues
Used as field type in:
Feature values in KeyValue format.
Feature values in proto Struct format.
The data key associated with this response. Will only be populated for [FeatureOnlineStoreService.StreamingFetchFeatureValues][] RPCs.
Response structure in the format of key (feature name) and (feature) value pair.
Used in:
List of feature names and values.
Feature name & value pair.
Used in:
Feature value.
Feature short name.
URI based data.
Used in:
Required. The IANA standard MIME type of the source data.
Required. URI.
RagFile status.
Used in:
Output only. RagFile state.
Output only. Only when the `state` field is ERROR.
RagFile state.
Used in:
RagFile state is unspecified.
RagFile resource has been created and indexed successfully.
RagFile resource is in a problematic state. See `error_message` field for details.
Assigns input data to training, validation, and test sets based on the given filters, data pieces not matched by any filter are ignored. Currently only supported for Datasets containing DataItems. If any of the filters in this message are to match nothing, then they can be set as '-' (the minus sign). Supported only for unstructured Datasets.
Used in:
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to train the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to validate the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
Required. A filter on DataItems of the Dataset. DataItems that match this filter are used to test the Model. A filter with same syntax as the one used in [DatasetService.ListDataItems][google.cloud.aiplatform.v1.DatasetService.ListDataItems] may be used. If a single DataItem is matched by more than one of the FilterSplit filters, then it is assigned to the first set that applies to it in the training, validation, test order.
A query to find a number of the nearest neighbors (most similar vectors) of a vector.
Used in:
Optional. Represents RRF algorithm that combines search results.
Required. The datapoint/vector whose nearest neighbors should be searched for.
The number of nearest neighbors to be retrieved from database for each query. If not set, will use the default from the service configuration (https://cloud.google.com/vertex-ai/docs/matching-engine/configuring-indexes#nearest-neighbor-search-config).
Crowding is a constraint on a neighbor list produced by nearest neighbor search requiring that no more than some value k' of the k neighbors returned have the same value of crowding_attribute. It's used for improving result diversity. This field is the maximum number of matches with the same crowding tag.
The number of neighbors to find via approximate search before exact reordering is performed. If not set, the default value from scam config is used; if set, this value must be > 0.
The fraction of the number of leaves to search, set at query time allows user to tune search performance. This value increase result in both search accuracy and latency increase. The value should be between 0.0 and 1.0. If not set or set to 0.0, query uses the default value specified in NearestNeighborSearchConfig.TreeAHConfig.fraction_leaf_nodes_to_search.
Parameters for RRF algorithm that combines search results.
Used in:
Required. Users can provide an alpha value to give more weight to dense vs sparse results. For example, if the alpha is 0, we only return sparse and if the alpha is 1, we only return dense.
Nearest neighbors for one query.
Used in:
The ID of the query datapoint.
All its neighbors.
A neighbor of the query vector.
Used in:
The datapoint of the neighbor. Note that full datapoints are returned only when "return_full_datapoint" is set to true. Otherwise, only the "datapoint_id" and "crowding_tag" fields are populated.
The distance between the neighbor and the dense embedding query.
The distance between the neighbor and the query sparse_embedding.
Input for fluency metric.
Used in:
Required. Spec for fluency score metric.
Required. Fluency instance.
Spec for fluency instance.
Used in:
Required. Output of the evaluated model.
Spec for fluency result.
Used in:
Output only. Fluency score.
Output only. Explanation for fluency score.
Output only. Confidence for fluency score.
Spec for fluency score metric.
Used in:
Optional. Which version to use for evaluation.
Assigns the input data to training, validation, and test sets as per the given fractions. Any of `training_fraction`, `validation_fraction` and `test_fraction` may optionally be provided, they must sum to up to 1. If the provided ones sum to less than 1, the remainder is assigned to sets as decided by Vertex AI. If none of the fractions are set, by default roughly 80% of data is used for training, 10% for validation, and 10% for test.
Used in:
The fraction of the input data that is to be used to train the Model.
The fraction of the input data that is to be used to validate the Model.
The fraction of the input data that is to be used to evaluate the Model.
Input for fulfillment metric.
Used in:
Required. Spec for fulfillment score metric.
Required. Fulfillment instance.
Spec for fulfillment instance.
Used in:
Required. Output of the evaluated model.
Required. Inference instruction prompt to compare prediction with.
Spec for fulfillment result.
Used in:
Output only. Fulfillment score.
Output only. Explanation for fulfillment score.
Output only. Confidence for fulfillment score.
Spec for fulfillment metric.
Used in:
Optional. Which version to use for evaluation.
A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values.
Used in:
Required. The name of the function to call. Matches [FunctionDeclaration.name].
Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details.
Function calling config.
Used in:
Optional. Function calling mode.
Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided.
Function calling mode.
Used in:
Unspecified function calling mode. This value should not be used.
Default model behavior, model decides to predict either function calls or natural language response.
Model is constrained to always predicting function calls only. If "allowed_function_names" are set, the predicted function calls will be limited to any one of "allowed_function_names", else the predicted function calls will be any one of the provided "function_declarations".
Model will not predict any function calls. Model behavior is same as when not passing any function declarations.
Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client.
Used in:
Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64.
Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function.
Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1
Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function.
The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction.
Used in:
Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name].
Required. The function response in JSON object format. Use "output" key to specify function output and "error" key to specify error details (if any). If "output" and "error" keys are not specified, then whole "response" is treated as function output.
The Google Cloud Storage location where the output is to be written to.
Used in:
, , , , , , , , , ,Required. Google Cloud Storage URI to output directory. If the uri doesn't end with '/', a '/' will be automatically appended. The directory is created if it doesn't exist.
The Google Cloud Storage location for the input content.
Used in:
, , , , , , ,Required. Google Cloud Storage URI(-s) to the input file(s). May contain wildcards. For more information on wildcards, see https://cloud.google.com/storage/docs/wildcards.
Request message for [PredictionService.GenerateContent].
Used as request type in: PredictionService.GenerateContent, PredictionService.StreamGenerateContent
Required. The fully qualified name of the publisher model or tuned model endpoint to use. Publisher model format: `projects/{project}/locations/{location}/publishers/*/models/*` Tuned model endpoint format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request.
Optional. The user provided system instructions for the model. Note: only text should be used in parts and content in each part will be in a separate paragraph.
Optional. The name of the cached content used as context to serve the prediction. Note: only used in explicit caching, where users can have control over caching (e.g. what content to cache) and enjoy guaranteed cost savings. Format: `projects/{project}/locations/{location}/cachedContents/{cachedContent}`
Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model.
Optional. Tool config. This config is shared for all tools provided in the request.
Optional. The labels with user-defined metadata for the request. It is used for billing and reporting only. Label keys and values can be no longer than 63 characters (Unicode codepoints) and can only contain lowercase letters, numeric characters, underscores, and dashes. International characters are allowed. Label values are optional. Label keys must start with a letter.
Optional. Per request settings for blocking unsafe content. Enforced on GenerateContentResponse.candidates.
Optional. Generation config.
Response message for [PredictionService.GenerateContent].
Used as response type in: PredictionService.GenerateContent, PredictionService.StreamGenerateContent
Output only. Generated candidates.
Output only. The model version used to generate the response.
Output only. Timestamp when the request is made to the server.
Output only. response_id is used to identify each response. It is the encoding of the event_id.
Output only. Content filter results for a prompt sent in the request. Note: Sent only in the first stream chunk. Only happens when no candidates were generated due to content violations.
Usage metadata about the response(s).
Content filter results for a prompt sent in the request.
Used in:
Output only. Blocked reason.
Output only. Safety ratings.
Output only. A readable block reason message.
Blocked reason enumeration.
Used in:
Unspecified blocked reason.
Candidates blocked due to safety.
Candidates blocked due to other reason.
Candidates blocked due to the terms which are included from the terminology blocklist.
Candidates blocked due to prohibited content.
Usage metadata about response(s).
Used in:
Number of tokens in the request. When `cached_content` is set, this is still the total effective prompt size meaning this includes the number of tokens in the cached content.
Number of tokens in the response(s).
Total token count for prompt and response candidates.
Output only. Number of tokens in the cached part in the input (the cached content).
Output only. List of modalities that were processed in the request input.
Output only. List of modalities of the cached content in the request input.
Output only. List of modalities that were returned in the response.
Generation config.
Used in:
,Optional. Controls the randomness of predictions.
Optional. If specified, nucleus sampling will be used.
Optional. If specified, top-k sampling will be used.
Optional. Number of candidates to generate.
Optional. The maximum number of output tokens to generate per message.
Optional. Stop sequences.
Optional. If true, export the logprobs results in response.
Optional. Logit probabilities.
Optional. Positive penalties.
Optional. Frequency penalties.
Optional. Seed.
Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature.
Optional. The `Schema` object allows the definition of input and output data types. These types can be objects, but also primitives and arrays. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). If set, a compatible response_mime_type must also be set. Compatible mimetypes: `application/json`: Schema for JSON response.
Optional. Routing configuration.
Optional. Config for thinking features. An error will be returned if this field is set for models that don't support thinking.
The configuration for routing the request to a specific model.
Used in:
Routing mode.
Automated routing.
Manual routing.
When automated routing is specified, the routing will be determined by the pretrained routing model and customer provided model routing preference.
Used in:
The model routing preference.
The model routing preference.
Used in:
Unspecified model routing preference.
Prefer higher quality over low cost.
Balanced model routing preference.
Prefer lower cost over higher quality.
When manual routing is set, the specified model will be used directly.
Used in:
The model name to use. Only the public LLM models are accepted. e.g. 'gemini-1.5-pro-001'.
Config for thinking features.
Used in:
Optional. Indicates the thinking budget in tokens. This is only applied when enable_thinking is true.
Generic Metadata shared by all operations.
Used in:
, , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,Output only. Partial failures encountered. E.g. single files that couldn't be read. This field should never exceed 20 entries. Status details field will contain standard Google Cloud error details.
Output only. Time when the operation was created.
Output only. Time when the operation was updated for the last time. If the operation has finished (successfully or not), this is the finish time.
Contains information about the source of the models generated from Generative AI Studio.
Used in:
Required. The public base model URI.
Request message for [FeaturestoreService.GetFeature][google.cloud.aiplatform.v1.FeaturestoreService.GetFeature]. Request message for [FeatureRegistryService.GetFeature][google.cloud.aiplatform.v1.FeatureRegistryService.GetFeature].
Used as request type in: FeatureRegistryService.GetFeature, FeaturestoreService.GetFeature
Required. The name of the Feature resource. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}`
The Google Drive location for the input content.
Used in:
,Required. Google Drive resource IDs.
The type and ID of the Google Drive resource.
Used in:
Required. The type of the Google Drive resource.
Required. The ID of the Google Drive resource.
The type of the Google Drive resource.
Used in:
Unspecified resource type.
File resource type.
Folder resource type.
Tool to retrieve public web data for grounding, powered by Google.
Used in:
Specifies the dynamic retrieval configuration for the given source.
Input for groundedness metric.
Used in:
Required. Spec for groundedness metric.
Required. Groundedness instance.
Spec for groundedness instance.
Used in:
Required. Output of the evaluated model.
Required. Background information provided in context used to compare against the prediction.
Spec for groundedness result.
Used in:
Output only. Groundedness score.
Output only. Explanation for groundedness score.
Output only. Confidence for groundedness score.
Spec for groundedness metric.
Used in:
Optional. Which version to use for evaluation.
Grounding chunk.
Used in:
Chunk type.
Grounding chunk from the web.
Grounding chunk from context retrieved by the retrieval tools.
Chunk from context retrieved by the retrieval tools.
Used in:
Tool-specific details about the retrieved context.
Additional context for the RAG retrieval result. This is only populated when using the RAG retrieval tool.
URI reference of the attribution.
Title of the attribution.
Text of the attribution.
Chunk from the web.
Used in:
URI reference of the chunk.
Title of the chunk.
Metadata returned to client when grounding is enabled.
Used in:
Optional. Web search queries for the following-up web search.
Optional. Google search entry for the following-up web searches.
List of supporting references retrieved from specified grounding source.
Optional. List of grounding support.
Optional. Output only. Retrieval metadata.
Grounding support.
Used in:
Segment of the content this support belongs to.
A list of indices (into 'grounding_chunk') specifying the citations associated with the claim. For instance [1,3,4] means that grounding_chunk[1], grounding_chunk[3], grounding_chunk[4] are the retrieved content attributed to the claim.
Confidence score of the support references. Ranges from 0 to 1. 1 is the most confident. This list must have the same size as the grounding_chunk_indices.
Harm categories that will block the content.
Used in:
,The harm category is unspecified.
The harm category is hate speech.
The harm category is dangerous content.
The harm category is harassment.
The harm category is sexually explicit content.
Deprecated: Election filter is not longer supported. The harm category is civic integrity.
Represents a HyperparameterTuningJob. A HyperparameterTuningJob has a Study specification and multiple CustomJobs with identical CustomJob specification.
Used as response type in: JobService.CreateHyperparameterTuningJob, JobService.GetHyperparameterTuningJob
Used as field type in:
,Output only. Resource name of the HyperparameterTuningJob.
Required. The display name of the HyperparameterTuningJob. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Required. Study configuration of the HyperparameterTuningJob.
Required. The desired total number of Trials.
Required. The desired number of Trials to run in parallel.
The number of failed Trials that need to be seen before failing the HyperparameterTuningJob. If set to 0, Vertex AI decides how many Trials must fail before the whole job fails.
Required. The spec of a trial job. The same spec applies to the CustomJobs created in all the trials.
Output only. Trials of the HyperparameterTuningJob.
Output only. The detailed state of the job.
Output only. Time when the HyperparameterTuningJob was created.
Output only. Time when the HyperparameterTuningJob for the first time entered the `JOB_STATE_RUNNING` state.
Output only. Time when the HyperparameterTuningJob entered any of the following states: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`.
Output only. Time when the HyperparameterTuningJob was most recently updated.
Output only. Only populated when job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED.
The labels with user-defined metadata to organize HyperparameterTuningJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Customer-managed encryption key options for a HyperparameterTuningJob. If this is set, then all resources created by the HyperparameterTuningJob will be encrypted with the provided encryption key.
Output only. Reserved for future use.
Output only. Reserved for future use.
Matcher for Features of an EntityType by Feature ID.
Used in:
Required. The following are accepted as `ids`: * A single-element list containing only `*`, which selects all Features in the target EntityType, or * A list containing only Feature IDs, which selects only Features with those IDs in the target EntityType.
Describes the location from where we import data into a Dataset, together with the labels that will be applied to the DataItems and the Annotations.
Used in:
The source of the input.
The Google Cloud Storage location for the input content.
Labels that will be applied to newly imported DataItems. If an identical DataItem as one being imported already exists in the Dataset, then these labels will be appended to these of the already existing one, and if labels with identical key is imported before, the old label value will be overwritten. If two DataItems are identical in the same import data operation, the labels will be combined and if key collision happens in this case, one of the values will be picked randomly. Two DataItems are considered identical if their content bytes are identical (e.g. image bytes or pdf bytes). These labels will be overridden by Annotation labels specified inside index file referenced by [import_schema_uri][google.cloud.aiplatform.v1.ImportDataConfig.import_schema_uri], e.g. jsonl file.
Labels that will be applied to newly imported Annotations. If two Annotations are identical, one of them will be deduped. Two Annotations are considered identical if their [payload][google.cloud.aiplatform.v1.Annotation.payload], [payload_schema_uri][google.cloud.aiplatform.v1.Annotation.payload_schema_uri] and all of their [labels][google.cloud.aiplatform.v1.Annotation.labels] are the same. These labels will be overridden by Annotation labels specified inside index file referenced by [import_schema_uri][google.cloud.aiplatform.v1.ImportDataConfig.import_schema_uri], e.g. jsonl file.
Required. Points to a YAML file stored on Google Cloud Storage describing the import format. Validation will be done against the schema. The schema is defined as an [OpenAPI 3.0.2 Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
Runtime operation information for [DatasetService.ImportData][google.cloud.aiplatform.v1.DatasetService.ImportData].
The common part of the operation metadata.
Response message for [DatasetService.ImportData][google.cloud.aiplatform.v1.DatasetService.ImportData].
(message has no fields)
Details of operations that perform import Feature values.
Operation metadata for Featurestore import Feature values.
Number of entities that have been imported by the operation.
Number of Feature values that have been imported by the operation.
The source URI from where Feature values are imported.
The number of rows in input source that weren't imported due to either * Not having any featureValues. * Having a null entityId. * Having a null timestamp. * Not being parsable (applicable for CSV sources).
The number rows that weren't ingested due to having timestamps outside the retention boundary.
List of ImportFeatureValues operations running under a single EntityType that are blocking this operation.
Defines the Feature value(s) to import.
Used in:
Required. ID of the Feature to import values of. This Feature must exist in the target EntityType, or the request will fail.
Source column to get the Feature values from. If not set, uses the column with the same name as the Feature ID.
Response message for [FeaturestoreService.ImportFeatureValues][google.cloud.aiplatform.v1.FeaturestoreService.ImportFeatureValues].
Number of entities that have been imported by the operation.
Number of Feature values that have been imported by the operation.
The number of rows in input source that weren't imported due to either * Not having any featureValues. * Having a null entityId. * Having a null timestamp. * Not being parsable (applicable for CSV sources).
The number rows that weren't ingested due to having feature timestamps outside the retention boundary.
Config for importing RagFiles.
Used in:
,The source of the import.
Google Cloud Storage location. Supports importing individual files as well as entire Google Cloud Storage directories. Sample formats: - `gs://bucket_name/my_directory/object_name/my_file.txt` - `gs://bucket_name/my_directory`
Google Drive location. Supports importing individual files as well as Google Drive folders.
Slack channels with their corresponding access tokens.
Jira queries with their corresponding authentication.
SharePoint sources.
Optional. If provided, all partial failures are written to the sink. Deprecated. Prefer to use the `import_result_sink`.
The Cloud Storage path to write partial failures to. Deprecated. Prefer to use `import_result_gcs_sink`.
The BigQuery destination to write partial failures to. It should be a bigquery table resource name (e.g. "bq://projectId.bqDatasetId.bqTableId"). The dataset must exist. If the table does not exist, it will be created with the expected schema. If the table exists, the schema will be validated and data will be added to this existing table. Deprecated. Prefer to use `import_result_bq_sink`.
Optional. If provided, all successfully imported files and all partial failures are written to the sink.
The Cloud Storage path to write import result to.
The BigQuery destination to write import result to. It should be a bigquery table resource name (e.g. "bq://projectId.bqDatasetId.bqTableId"). The dataset must exist. If the table does not exist, it will be created with the expected schema. If the table exists, the schema will be validated and data will be added to this existing table.
Specifies the transformation config for RagFiles.
Optional. Specifies the parsing config for RagFiles. RAG will use the default parser if this field is not set.
Optional. The max number of queries per minute that this job is allowed to make to the embedding model specified on the corpus. This value is specific to this job and not shared across other import jobs. Consult the Quotas page on the project to set an appropriate value here. If unspecified, a default value of 1,000 QPM would be used.
Runtime operation information for [VertexRagDataService.ImportRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ImportRagFiles].
The operation generic information.
The resource ID of RagCorpus that this operation is executed on.
Output only. The config that was passed in the ImportRagFilesRequest.
The progress percentage of the operation. Value is in the range [0, 100]. This percentage is calculated as follows: progress_percentage = 100 * (successes + failures + skips) / total
Response message for [VertexRagDataService.ImportRagFiles][google.cloud.aiplatform.v1.VertexRagDataService.ImportRagFiles].
The location into which the partial failures were written.
The Google Cloud Storage path into which the partial failures were written.
The BigQuery table into which the partial failures were written.
The number of RagFiles that had been imported into the RagCorpus.
The number of RagFiles that had failed while importing into the RagCorpus.
The number of RagFiles that was skipped while importing into the RagCorpus.
A representation of a collection of database items organized in a way that allows for approximate nearest neighbor (a.k.a ANN) algorithms search.
Used as response type in: IndexService.GetIndex
Used as field type in:
, ,Output only. The resource name of the Index.
Required. The display name of the Index. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the Index.
Immutable. Points to a YAML file stored on Google Cloud Storage describing additional information about the Index, that is specific to it. Unset if the Index does not have any additional information. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
An additional information about the Index; the schema of the metadata can be found in [metadata_schema][google.cloud.aiplatform.v1.Index.metadata_schema_uri].
Output only. The pointers to DeployedIndexes created from this Index. An Index can be only deleted if all its DeployedIndexes had been undeployed first.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Indexes. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Output only. Timestamp when this Index was created.
Output only. Timestamp when this Index was most recently updated. This also includes any update to the contents of the Index. Note that Operations working on this Index may have their [Operations.metadata.generic_metadata.update_time] [google.cloud.aiplatform.v1.GenericOperationMetadata.update_time] a little after the value of this timestamp, yet that does not mean their results are not already reflected in the Index. Result of any successfully completed Operation on the Index is reflected in it.
Output only. Stats of the index resource.
Immutable. The update method to use with this Index. If not set, BATCH_UPDATE will be used by default.
Immutable. Customer-managed encryption key spec for an Index. If set, this Index and all sub-resources of this Index will be secured by this key.
Output only. Reserved for future use.
Output only. Reserved for future use.
The update method of an Index.
Used in:
Should not be used.
BatchUpdate: user can call UpdateIndex with files on Cloud Storage of Datapoints to update.
StreamUpdate: user can call UpsertDatapoints/DeleteDatapoints to update the Index and the updates will be applied in corresponding DeployedIndexes in nearly real-time.
A datapoint of Index.
Used in:
, , ,Required. Unique identifier of the datapoint.
Required. Feature embedding vector for dense index. An array of numbers with the length of [NearestNeighborSearchConfig.dimensions].
Optional. Feature embedding vector for sparse index.
Optional. List of Restrict of the datapoint, used to perform "restricted searches" where boolean rule are used to filter the subset of the database eligible for matching. This uses categorical tokens. See: https://cloud.google.com/vertex-ai/docs/matching-engine/filtering
Optional. List of Restrict of the datapoint, used to perform "restricted searches" where boolean rule are used to filter the subset of the database eligible for matching. This uses numeric comparisons.
Optional. CrowdingTag of the datapoint, the number of neighbors to return in each crowding can be configured during query.
Crowding tag is a constraint on a neighbor list produced by nearest neighbor search requiring that no more than some value k' of the k neighbors returned have the same value of crowding_attribute.
Used in:
The attribute value used for crowding. The maximum number of neighbors to return per crowding attribute value (per_crowding_attribute_num_neighbors) is configured per-query. This field is ignored if per_crowding_attribute_num_neighbors is larger than the total number of neighbors to return for a given query.
This field allows restricts to be based on numeric comparisons rather than categorical tokens.
Used in:
The type of Value must be consistent for all datapoints with a given namespace name. This is verified at runtime.
Represents 64 bit integer.
Represents 32 bit float.
Represents 64 bit float.
The namespace of this restriction. e.g.: cost.
This MUST be specified for queries and must NOT be specified for datapoints.
Which comparison operator to use. Should be specified for queries only; specifying this for a datapoint is an error. Datapoints for which Operator is true relative to the query's Value field will be allowlisted.
Used in:
Default value of the enum.
Datapoints are eligible iff their value is < the query's.
Datapoints are eligible iff their value is <= the query's.
Datapoints are eligible iff their value is == the query's.
Datapoints are eligible iff their value is >= the query's.
Datapoints are eligible iff their value is > the query's.
Datapoints are eligible iff their value is != the query's.
Restriction of a datapoint which describe its attributes(tokens) from each of several attribute categories(namespaces).
Used in:
The namespace of this restriction. e.g.: color.
The attributes to allow in this namespace. e.g.: 'red'
The attributes to deny in this namespace. e.g.: 'blue'
Feature embedding vector for sparse index. An array of numbers whose values are located in the specified dimensions.
Used in:
Required. The list of embedding values of the sparse vector.
Required. The list of indexes for the embedding values of the sparse vector.
Indexes are deployed into it. An IndexEndpoint can have multiple DeployedIndexes.
Used as response type in: IndexEndpointService.GetIndexEndpoint, IndexEndpointService.UpdateIndexEndpoint
Used as field type in:
, ,Output only. The resource name of the IndexEndpoint.
Required. The display name of the IndexEndpoint. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the IndexEndpoint.
Output only. The indexes deployed in this endpoint.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your IndexEndpoints. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Output only. Timestamp when this IndexEndpoint was created.
Output only. Timestamp when this IndexEndpoint was last updated. This timestamp is not updated when the endpoint's DeployedIndexes are updated, e.g. due to updates of the original Indexes they are the deployments of.
Optional. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks) to which the IndexEndpoint should be peered. Private services access must already be configured for the network. If left unspecified, the Endpoint is not peered with any network. [network][google.cloud.aiplatform.v1.IndexEndpoint.network] and [private_service_connect_config][google.cloud.aiplatform.v1.IndexEndpoint.private_service_connect_config] are mutually exclusive. [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name.
Optional. Deprecated: If true, expose the IndexEndpoint via private service connect. Only one of the fields, [network][google.cloud.aiplatform.v1.IndexEndpoint.network] or [enable_private_service_connect][google.cloud.aiplatform.v1.IndexEndpoint.enable_private_service_connect], can be set.
Optional. Configuration for private service connect. [network][google.cloud.aiplatform.v1.IndexEndpoint.network] and [private_service_connect_config][google.cloud.aiplatform.v1.IndexEndpoint.private_service_connect_config] are mutually exclusive.
Optional. If true, the deployed index will be accessible through public endpoint.
Output only. If [public_endpoint_enabled][google.cloud.aiplatform.v1.IndexEndpoint.public_endpoint_enabled] is true, this field will be populated with the domain name to use for this index endpoint.
Immutable. Customer-managed encryption key spec for an IndexEndpoint. If set, this IndexEndpoint and all sub-resources of this IndexEndpoint will be secured by this key.
Output only. Reserved for future use.
Output only. Reserved for future use.
IndexPrivateEndpoints proto is used to provide paths for users to send requests via private endpoints (e.g. private service access, private service connect). To send request via private service access, use match_grpc_address. To send request via private service connect, use service_attachment.
Used in:
Output only. The ip address used to send match gRPC requests.
Output only. The name of the service attachment resource. Populated if private service connect is enabled.
Output only. PscAutomatedEndpoints is populated if private service connect is enabled if PscAutomatedConfig is set.
Stats of the Index.
Used in:
Output only. The number of dense vectors in the Index.
Output only. The number of sparse vectors in the Index.
Output only. The number of shards in the Index.
Specifies Vertex AI owned input data to be used for training, and possibly evaluating, the Model.
Used in:
The instructions how the input data should be split between the training, validation and test sets. If no split type is provided, the [fraction_split][google.cloud.aiplatform.v1.InputDataConfig.fraction_split] is used by default.
Split based on fractions defining the size of each set.
Split based on the provided filters for each set.
Supported only for tabular Datasets. Split based on a predefined key.
Supported only for tabular Datasets. Split based on the timestamp of the input data pieces.
Supported only for tabular Datasets. Split based on the distribution of the specified column.
Only applicable to Custom and Hyperparameter Tuning TrainingPipelines. The destination of the training data to be written to. Supported destination file formats: * For non-tabular data: "jsonl". * For tabular data: "csv" and "bigquery". The following Vertex AI environment variables are passed to containers or python modules of the training task when this field is set: * AIP_DATA_FORMAT : Exported data format. * AIP_TRAINING_DATA_URI : Sharded exported training data uris. * AIP_VALIDATION_DATA_URI : Sharded exported validation data uris. * AIP_TEST_DATA_URI : Sharded exported test data uris.
The Cloud Storage location where the training data is to be written to. In the given directory a new directory is created with name: `dataset-<dataset-id>-<annotation-type>-<timestamp-of-training-call>` where timestamp is in YYYY-MM-DDThh:mm:ss.sssZ ISO-8601 format. All training input data is written into that directory. The Vertex AI environment variables representing Cloud Storage data URIs are represented in the Cloud Storage wildcard format to support sharded data. e.g.: "gs://.../training-*.jsonl" * AIP_DATA_FORMAT = "jsonl" for non-tabular data, "csv" for tabular data * AIP_TRAINING_DATA_URI = "gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/training-*.${AIP_DATA_FORMAT}" * AIP_VALIDATION_DATA_URI = "gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/validation-*.${AIP_DATA_FORMAT}" * AIP_TEST_DATA_URI = "gcs_destination/dataset-<dataset-id>-<annotation-type>-<time>/test-*.${AIP_DATA_FORMAT}"
Only applicable to custom training with tabular Dataset with BigQuery source. The BigQuery project location where the training data is to be written to. In the given project a new dataset is created with name `dataset_<dataset-id>_<annotation-type>_<timestamp-of-training-call>` where timestamp is in YYYY_MM_DDThh_mm_ss_sssZ format. All training input data is written into that dataset. In the dataset three tables are created, `training`, `validation` and `test`. * AIP_DATA_FORMAT = "bigquery". * AIP_TRAINING_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.training" * AIP_VALIDATION_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.validation" * AIP_TEST_DATA_URI = "bigquery_destination.dataset_<dataset-id>_<annotation-type>_<time>.test"
Required. The ID of the Dataset in the same Project and Location which data will be used to train the Model. The Dataset must use schema compatible with Model being trained, and what is compatible should be described in the used TrainingPipeline's [training_task_definition] [google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition]. For tabular Datasets, all their data is exported to training, to pick and choose from.
Applicable only to Datasets that have DataItems and Annotations. A filter on Annotations of the Dataset. Only Annotations that both match this filter and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on (for the auto-assigned that role is decided by Vertex AI). A filter with same syntax as the one used in [ListAnnotations][google.cloud.aiplatform.v1.DatasetService.ListAnnotations] may be used, but note here it filters across all Annotations of the Dataset, and not just within a single DataItem.
Applicable only to custom training with Datasets that have DataItems and Annotations. Cloud Storage URI that points to a YAML file describing the annotation schema. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). The schema files that can be used here are found in gs://google-cloud-aiplatform/schema/dataset/annotation/ , note that the chosen schema must be consistent with [metadata][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri] of the Dataset specified by [dataset_id][google.cloud.aiplatform.v1.InputDataConfig.dataset_id]. Only Annotations that both match this schema and belong to DataItems not ignored by the split method are used in respectively training, validation or test role, depending on the role of the DataItem they are on. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [annotations_filter][google.cloud.aiplatform.v1.InputDataConfig.annotations_filter] and [annotation_schema_uri][google.cloud.aiplatform.v1.InputDataConfig.annotation_schema_uri].
Only applicable to Datasets that have SavedQueries. The ID of a SavedQuery (annotation set) under the Dataset specified by [dataset_id][google.cloud.aiplatform.v1.InputDataConfig.dataset_id] used for filtering Annotations for training. Only Annotations that are associated with this SavedQuery are used in respectively training. When used in conjunction with [annotations_filter][google.cloud.aiplatform.v1.InputDataConfig.annotations_filter], the Annotations used for training are filtered by both [saved_query_id][google.cloud.aiplatform.v1.InputDataConfig.saved_query_id] and [annotations_filter][google.cloud.aiplatform.v1.InputDataConfig.annotations_filter]. Only one of [saved_query_id][google.cloud.aiplatform.v1.InputDataConfig.saved_query_id] and [annotation_schema_uri][google.cloud.aiplatform.v1.InputDataConfig.annotation_schema_uri] should be specified as both of them represent the same thing: problem type.
Whether to persist the ML use assignment to data item system labels.
A list of int64 values.
Used in:
A list of int64 values.
An attribution method that computes the Aumann-Shapley value taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1703.01365
Used in:
Required. The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is within the desired error range. Valid range of its value is [1, 100], inclusively.
Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
Config for IG with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383
The Jira source for the ImportRagFilesRequest.
Used in:
,Required. The Jira queries.
JiraQueries contains the Jira queries and corresponding authentication.
Used in:
A list of Jira projects to import in their entirety.
A list of custom Jira queries to import. For information about JQL (Jira Query Language), see https://support.atlassian.com/jira-service-management-cloud/docs/use-advanced-search-with-jira-query-language-jql/
Required. The Jira email address.
Required. The Jira server URI.
Required. The SecretManager secret version resource name (e.g. projects/{project}/secrets/{secret}/versions/{version}) storing the Jira API key. See [Manage API tokens for your Atlassian account](https://support.atlassian.com/atlassian-account/docs/manage-api-tokens-for-your-atlassian-account/).
Describes the state of a job.
Used in:
, , , , , , ,The job state is unspecified.
The job has been just created or resumed and processing has not yet begun.
The service is preparing to run the job.
The job is in progress.
The job completed successfully.
The job failed.
The job is being cancelled. From this state the job may only go to either `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
The job has been cancelled.
The job has been stopped, and can be resumed.
The job has expired.
The job is being updated. Only jobs in the `RUNNING` state can be updated. After updating, the job goes back to the `RUNNING` state.
The job is partially succeeded, some results may be missing due to errors.
Contains information about the Large Model.
Used in:
Required. The unique name of the large Foundation or pre-built model. Like "chat-bison", "text-bison". Or model name with version ID, like "chat-bison@001", "text-bison@005", etc.
A subgraph of the overall lineage graph. Event edges connect Artifact and Execution nodes.
Used as response type in: MetadataService.QueryArtifactLineageSubgraph, MetadataService.QueryContextLineageSubgraph, MetadataService.QueryExecutionInputsAndOutputs
The Artifact nodes in the subgraph.
The Execution nodes in the subgraph.
The Event edges between Artifacts and Executions in the subgraph.
Request message for [FeaturestoreService.ListFeatures][google.cloud.aiplatform.v1.FeaturestoreService.ListFeatures]. Request message for [FeatureRegistryService.ListFeatures][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatures].
Used as request type in: FeatureRegistryService.ListFeatures, FeaturestoreService.ListFeatures
Required. The resource name of the Location to list Features. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}`
Lists the Features that match the filter expression. The following filters are supported: * `value_type`: Supports = and != comparisons. * `create_time`: Supports =, !=, <, >, >=, and <= comparisons. Values must be in RFC 3339 format. * `update_time`: Supports =, !=, <, >, >=, and <= comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality as well as key presence. Examples: * `value_type = DOUBLE` --> Features whose type is DOUBLE. * `create_time > \"2020-01-31T15:30:00.000000Z\" OR update_time > \"2020-01-31T15:30:00.000000Z\"` --> EntityTypes created or updated after 2020-01-31T15:30:00.000000Z. * `labels.active = yes AND labels.env = prod` --> Features having both (active: yes) and (env: prod) labels. * `labels.env: *` --> Any Feature which has a label with 'env' as the key.
The maximum number of Features to return. The service may return fewer than this value. If unspecified, at most 1000 Features will be returned. The maximum value is 1000; any value greater than 1000 will be coerced to 1000.
A page token, received from a previous [FeaturestoreService.ListFeatures][google.cloud.aiplatform.v1.FeaturestoreService.ListFeatures] call or [FeatureRegistryService.ListFeatures][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatures] call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to [FeaturestoreService.ListFeatures][google.cloud.aiplatform.v1.FeaturestoreService.ListFeatures] or [FeatureRegistryService.ListFeatures][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatures] must match the call that provided the page token.
A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported fields: * `feature_id` * `value_type` (Not supported for FeatureRegistry Feature) * `create_time` * `update_time`
Mask specifying which fields to read.
Only applicable for Vertex AI Feature Store (Legacy). If set, return the most recent [ListFeaturesRequest.latest_stats_count][google.cloud.aiplatform.v1.ListFeaturesRequest.latest_stats_count] of stats for each Feature in response. Valid value is [0, 10]. If number of stats exists < [ListFeaturesRequest.latest_stats_count][google.cloud.aiplatform.v1.ListFeaturesRequest.latest_stats_count], return all existing stats.
Response message for [FeaturestoreService.ListFeatures][google.cloud.aiplatform.v1.FeaturestoreService.ListFeatures]. Response message for [FeatureRegistryService.ListFeatures][google.cloud.aiplatform.v1.FeatureRegistryService.ListFeatures].
Used as response type in: FeatureRegistryService.ListFeatures, FeaturestoreService.ListFeatures
The Features matching the request.
A token, which can be sent as [ListFeaturesRequest.page_token][google.cloud.aiplatform.v1.ListFeaturesRequest.page_token] to retrieve the next page. If this field is omitted, there are no subsequent pages.
Logprobs Result
Used in:
Length = total number of decoding steps.
Length = total number of decoding steps. The chosen candidates may or may not be in top_candidates.
Candidate for the logprobs token and score.
Used in:
,The candidate’s token string value.
The candidate’s token id value.
The candidate's log probability.
Candidates with top log probabilities at each decoding step.
Used in:
Sorted by log probability in descending order.
Specification of a single machine.
Used in:
, , , , , ,Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For [DeployedModel][google.cloud.aiplatform.v1.DeployedModel] this field is optional, and the default value is `n1-standard-2`. For [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob] or as part of [WorkerPoolSpec][google.cloud.aiplatform.v1.WorkerPoolSpec] this field is required.
Immutable. The type of accelerator(s) that may be attached to the machine as per [accelerator_count][google.cloud.aiplatform.v1.MachineSpec.accelerator_count].
The number of accelerators to attach to the machine.
Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1").
Optional. Immutable. Configuration controlling how this resource pool consumes reservation.
Manual batch tuning parameters.
Used in:
Immutable. The number of the records (e.g. instances) of the operation given in each batch to a machine replica. Machine type, and size of a single record should be considered when setting this parameter, higher value speeds up the batch operation's execution, but too high value will result in a whole batch not fitting in a machine's memory, and the whole operation will fail. The default value is 64.
A message representing a Measurement of a Trial. A Measurement contains the Metrics got by executing a Trial using suggested hyperparameter values.
Used in:
, , ,Output only. Time that the Trial has been running at the point of this Measurement.
Output only. The number of steps the machine learning model has been trained for. Must be non-negative.
Output only. A list of metrics got by evaluating the objective functions using suggested Parameter values.
A message representing a metric in the measurement.
Used in:
Output only. The ID of the Metric. The Metric should be defined in [StudySpec's Metrics][google.cloud.aiplatform.v1.StudySpec.metrics].
Output only. The value for this metric.
Instance of a general MetadataSchema.
Used as response type in: MetadataService.CreateMetadataSchema, MetadataService.GetMetadataSchema
Used as field type in:
,Output only. The resource name of the MetadataSchema.
The version of the MetadataSchema. The version's format must match the following regular expression: `^[0-9]+[.][0-9]+[.][0-9]+$`, which would allow to order/compare different versions. Example: 1.0.0, 1.0.1, etc.
Required. The raw YAML string representation of the MetadataSchema. The combination of [MetadataSchema.version] and the schema name given by `title` in [MetadataSchema.schema] must be unique within a MetadataStore. The schema is defined as an OpenAPI 3.0.2 [MetadataSchema Object](https://github.com/OAI/OpenAPI-Specification/blob/master/versions/3.0.2.md#schemaObject)
The type of the MetadataSchema. This is a property that identifies which metadata types will use the MetadataSchema.
Output only. Timestamp when this MetadataSchema was created.
Description of the Metadata Schema
Describes the type of the MetadataSchema.
Used in:
Unspecified type for the MetadataSchema.
A type indicating that the MetadataSchema will be used by Artifacts.
A typee indicating that the MetadataSchema will be used by Executions.
A state indicating that the MetadataSchema will be used by Contexts.
Instance of a metadata store. Contains a set of metadata that can be queried.
Used as response type in: MetadataService.GetMetadataStore
Used as field type in:
,Output only. The resource name of the MetadataStore instance.
Output only. Timestamp when this MetadataStore was created.
Output only. Timestamp when this MetadataStore was last updated.
Customer-managed encryption key spec for a Metadata Store. If set, this Metadata Store and all sub-resources of this Metadata Store are secured using this key.
Description of the MetadataStore.
Output only. State information of the MetadataStore.
Optional. Dataplex integration settings.
Represents Dataplex integration settings.
Used in:
Optional. Whether or not Data Lineage synchronization is enabled for Vertex Pipelines.
Represents state information for a MetadataStore.
Used in:
The disk utilization of the MetadataStore in bytes.
Input for MetricX metric.
Used in:
Required. Spec for Metricx metric.
Required. Metricx instance.
Spec for MetricX instance - The fields used for evaluation are dependent on the MetricX version.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Optional. Source text in original language.
Spec for MetricX result - calculates the MetricX score for the given instance using the version specified in the spec.
Used in:
Output only. MetricX score. Range depends on version.
Spec for MetricX metric.
Used in:
Required. Which version to use for evaluation.
Optional. Source language in BCP-47 format.
Optional. Target language in BCP-47 format. Covers both prediction and reference.
MetricX Version options.
Used in:
MetricX version unspecified.
MetricX 2024 (2.6) for translation + reference (reference-based).
MetricX 2024 (2.6) for translation + source (QE).
MetricX 2024 (2.6) for translation + source + reference (source-reference-combined).
Represents one resource that exists in automl.googleapis.com, datalabeling.googleapis.com or ml.googleapis.com.
Used in:
,Output only. Represents one Version in ml.googleapis.com.
Output only. Represents one Model in automl.googleapis.com.
Output only. Represents one Dataset in automl.googleapis.com.
Output only. Represents one Dataset in datalabeling.googleapis.com.
Output only. Timestamp when the last migration attempt on this MigratableResource started. Will not be set if there's no migration attempt on this MigratableResource.
Output only. Timestamp when this MigratableResource was last updated.
Represents one Dataset in automl.googleapis.com.
Used in:
Full resource name of automl Dataset. Format: `projects/{project}/locations/{location}/datasets/{dataset}`.
The Dataset's display name in automl.googleapis.com.
Represents one Model in automl.googleapis.com.
Used in:
Full resource name of automl Model. Format: `projects/{project}/locations/{location}/models/{model}`.
The Model's display name in automl.googleapis.com.
Represents one Dataset in datalabeling.googleapis.com.
Used in:
Full resource name of data labeling Dataset. Format: `projects/{project}/datasets/{dataset}`.
The Dataset's display name in datalabeling.googleapis.com.
The migratable AnnotatedDataset in datalabeling.googleapis.com belongs to the data labeling Dataset.
Represents one AnnotatedDataset in datalabeling.googleapis.com.
Used in:
Full resource name of data labeling AnnotatedDataset. Format: `projects/{project}/datasets/{dataset}/annotatedDatasets/{annotated_dataset}`.
The AnnotatedDataset's display name in datalabeling.googleapis.com.
Represents one model Version in ml.googleapis.com.
Used in:
The ml.googleapis.com endpoint that this model Version currently lives in. Example values: * ml.googleapis.com * us-centrall-ml.googleapis.com * europe-west4-ml.googleapis.com * asia-east1-ml.googleapis.com
Full resource name of ml engine model Version. Format: `projects/{project}/models/{model}/versions/{version}`.
Config of migrating one resource from automl.googleapis.com, datalabeling.googleapis.com and ml.googleapis.com to Vertex AI.
Used in:
,Config for migrating Version in ml.googleapis.com to Vertex AI's Model.
Config for migrating Model in automl.googleapis.com to Vertex AI's Model.
Config for migrating Dataset in automl.googleapis.com to Vertex AI's Dataset.
Config for migrating Dataset in datalabeling.googleapis.com to Vertex AI's Dataset.
Config for migrating Dataset in automl.googleapis.com to Vertex AI's Dataset.
Used in:
Required. Full resource name of automl Dataset. Format: `projects/{project}/locations/{location}/datasets/{dataset}`.
Required. Display name of the Dataset in Vertex AI. System will pick a display name if unspecified.
Config for migrating Model in automl.googleapis.com to Vertex AI's Model.
Used in:
Required. Full resource name of automl Model. Format: `projects/{project}/locations/{location}/models/{model}`.
Optional. Display name of the model in Vertex AI. System will pick a display name if unspecified.
Config for migrating Dataset in datalabeling.googleapis.com to Vertex AI's Dataset.
Used in:
Required. Full resource name of data labeling Dataset. Format: `projects/{project}/datasets/{dataset}`.
Optional. Display name of the Dataset in Vertex AI. System will pick a display name if unspecified.
Optional. Configs for migrating AnnotatedDataset in datalabeling.googleapis.com to Vertex AI's SavedQuery. The specified AnnotatedDatasets have to belong to the datalabeling Dataset.
Config for migrating AnnotatedDataset in datalabeling.googleapis.com to Vertex AI's SavedQuery.
Used in:
Required. Full resource name of data labeling AnnotatedDataset. Format: `projects/{project}/datasets/{dataset}/annotatedDatasets/{annotated_dataset}`.
Config for migrating version in ml.googleapis.com to Vertex AI's Model.
Used in:
Required. The ml.googleapis.com endpoint that this model version should be migrated from. Example values: * ml.googleapis.com * us-centrall-ml.googleapis.com * europe-west4-ml.googleapis.com * asia-east1-ml.googleapis.com
Required. Full resource name of ml engine model version. Format: `projects/{project}/models/{model}/versions/{version}`.
Required. Display name of the model in Vertex AI. System will pick a display name if unspecified.
Describes a successfully migrated resource.
Used in:
After migration, the resource name in Vertex AI.
Migrated Dataset's resource name.
Migrated Model's resource name.
Before migration, the identifier in ml.googleapis.com, automl.googleapis.com or datalabeling.googleapis.com.
Content Part modality
Used in:
Unspecified modality.
Plain text.
Image.
Video.
Audio.
Document, e.g. PDF.
Represents token counting info for a single modality.
Used in:
,The modality associated with this token count.
Number of tokens.
A trained machine learning Model.
Used as response type in: ModelService.GetModel, ModelService.MergeVersionAliases, ModelService.UpdateModel
Used as field type in:
, , , ,The resource name of the Model.
Output only. Immutable. The version ID of the model. A new version is committed when a new model version is uploaded or trained under an existing model id. It is an auto-incrementing decimal number in string representation.
User provided version aliases so that a model version can be referenced via alias (i.e. `projects/{project}/locations/{location}/models/{model_id}@{version_alias}` instead of auto-generated version id (i.e. `projects/{project}/locations/{location}/models/{model_id}@{version_id})`. The format is [a-z][a-zA-Z0-9-]{0,126}[a-z0-9] to distinguish from version_id. A default version alias will be created for the first version of the model, and there must be exactly one default version alias for a model.
Output only. Timestamp when this version was created.
Output only. Timestamp when this version was most recently updated.
Required. The display name of the Model. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the Model.
The description of this version.
The default checkpoint id of a model version.
The schemata that describe formats of the Model's predictions and explanations as given and returned via [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict] and [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain].
Immutable. Points to a YAML file stored on Google Cloud Storage describing additional information about the Model, that is specific to it. Unset if the Model does not have any additional information. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). AutoML Models always have this field populated by Vertex AI, if no additional metadata is needed, this field is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Immutable. An additional information about the Model; the schema of the metadata can be found in [metadata_schema][google.cloud.aiplatform.v1.Model.metadata_schema_uri]. Unset if the Model does not have any additional information.
Output only. The formats in which this Model may be exported. If empty, this Model is not available for export.
Output only. The resource name of the TrainingPipeline that uploaded this Model, if any.
Optional. This field is populated if the model is produced by a pipeline job.
Input only. The specification of the container that is to be used when deploying this Model. The specification is ingested upon [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel], and all binaries it contains are copied and stored internally by Vertex AI. Not required for AutoML Models.
Immutable. The path to the directory containing the Model artifact and any of its supporting files. Not required for AutoML Models.
Output only. When this Model is deployed, its prediction resources are described by the `prediction_resources` field of the [Endpoint.deployed_models][google.cloud.aiplatform.v1.Endpoint.deployed_models] object. Because not all Models support all resource configuration types, the configuration types this Model supports are listed here. If no configuration types are listed, the Model cannot be deployed to an [Endpoint][google.cloud.aiplatform.v1.Endpoint] and does not support online predictions ([PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict] or [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain]). Such a Model can serve predictions by using a [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob], if it has at least one entry each in [supported_input_storage_formats][google.cloud.aiplatform.v1.Model.supported_input_storage_formats] and [supported_output_storage_formats][google.cloud.aiplatform.v1.Model.supported_output_storage_formats].
Output only. The formats this Model supports in [BatchPredictionJob.input_config][google.cloud.aiplatform.v1.BatchPredictionJob.input_config]. If [PredictSchemata.instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] exists, the instances should be given as per that schema. The possible formats are: * `jsonl` The JSON Lines format, where each instance is a single line. Uses [GcsSource][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.gcs_source]. * `csv` The CSV format, where each instance is a single comma-separated line. The first line in the file is the header, containing comma-separated field names. Uses [GcsSource][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.gcs_source]. * `tf-record` The TFRecord format, where each instance is a single record in tfrecord syntax. Uses [GcsSource][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.gcs_source]. * `tf-record-gzip` Similar to `tf-record`, but the file is gzipped. Uses [GcsSource][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.gcs_source]. * `bigquery` Each instance is a single row in BigQuery. Uses [BigQuerySource][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig.bigquery_source]. * `file-list` Each line of the file is the location of an instance to process, uses `gcs_source` field of the [InputConfig][google.cloud.aiplatform.v1.BatchPredictionJob.InputConfig] object. If this Model doesn't support any of these formats it means it cannot be used with a [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob]. However, if it has [supported_deployment_resources_types][google.cloud.aiplatform.v1.Model.supported_deployment_resources_types], it could serve online predictions by using [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict] or [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain].
Output only. The formats this Model supports in [BatchPredictionJob.output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config]. If both [PredictSchemata.instance_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.instance_schema_uri] and [PredictSchemata.prediction_schema_uri][google.cloud.aiplatform.v1.PredictSchemata.prediction_schema_uri] exist, the predictions are returned together with their instances. In other words, the prediction has the original instance data first, followed by the actual prediction content (as per the schema). The possible formats are: * `jsonl` The JSON Lines format, where each prediction is a single line. Uses [GcsDestination][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.gcs_destination]. * `csv` The CSV format, where each prediction is a single comma-separated line. The first line in the file is the header, containing comma-separated field names. Uses [GcsDestination][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.gcs_destination]. * `bigquery` Each prediction is a single row in a BigQuery table, uses [BigQueryDestination][google.cloud.aiplatform.v1.BatchPredictionJob.OutputConfig.bigquery_destination] . If this Model doesn't support any of these formats it means it cannot be used with a [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob]. However, if it has [supported_deployment_resources_types][google.cloud.aiplatform.v1.Model.supported_deployment_resources_types], it could serve online predictions by using [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict] or [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain].
Output only. Timestamp when this Model was uploaded into Vertex AI.
Output only. Timestamp when this Model was most recently updated.
Output only. The pointers to DeployedModels created from this Model. Note that Model could have been deployed to Endpoints in different Locations.
The default explanation specification for this Model. The Model can be used for [requesting explanation][google.cloud.aiplatform.v1.PredictionService.Explain] after being [deployed][google.cloud.aiplatform.v1.EndpointService.DeployModel] if it is populated. The Model can be used for [batch explanation][google.cloud.aiplatform.v1.BatchPredictionJob.generate_explanation] if it is populated. All fields of the explanation_spec can be overridden by [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] of [DeployModelRequest.deployed_model][google.cloud.aiplatform.v1.DeployModelRequest.deployed_model], or [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] of [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob]. If the default explanation specification is not set for this Model, this Model can still be used for [requesting explanation][google.cloud.aiplatform.v1.PredictionService.Explain] by setting [explanation_spec][google.cloud.aiplatform.v1.DeployedModel.explanation_spec] of [DeployModelRequest.deployed_model][google.cloud.aiplatform.v1.DeployModelRequest.deployed_model] and for [batch explanation][google.cloud.aiplatform.v1.BatchPredictionJob.generate_explanation] by setting [explanation_spec][google.cloud.aiplatform.v1.BatchPredictionJob.explanation_spec] of [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob].
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize your Models. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Stats of data used for training or evaluating the Model. Only populated when the Model is trained by a TrainingPipeline with [data_input_config][google.cloud.aiplatform.v1.TrainingPipeline.input_data_config].
Customer-managed encryption key spec for a Model. If set, this Model and all sub-resources of this Model will be secured by this key.
Output only. Source of a model. It can either be automl training pipeline, custom training pipeline, BigQuery ML, or saved and tuned from Genie or Model Garden.
Output only. If this Model is a copy of another Model, this contains info about the original.
Output only. The resource name of the Artifact that was created in MetadataStore when creating the Model. The Artifact resource name pattern is `projects/{project}/locations/{location}/metadataStores/{metadata_store}/artifacts/{artifact}`.
Optional. User input field to specify the base model source. Currently it only supports specifing the Model Garden models and Genie models.
Output only. Reserved for future use.
Output only. Reserved for future use.
Optional. Output only. The checkpoints of the model.
User input field to specify the base model source. Currently it only supports specifing the Model Garden models and Genie models.
Used in:
Source information of Model Garden models.
Information about the base model of Genie models.
Stats of data used for train or evaluate the Model.
Used in:
,Number of DataItems that were used for training this Model.
Number of DataItems that were used for validating this Model during training.
Number of DataItems that were used for evaluating this Model. If the Model is evaluated multiple times, this will be the number of test DataItems used by the first evaluation. If the Model is not evaluated, the number is 0.
Number of Annotations that are used for training this Model.
Number of Annotations that are used for validating this Model during training.
Number of Annotations that are used for evaluating this Model. If the Model is evaluated multiple times, this will be the number of test Annotations used by the first evaluation. If the Model is not evaluated, the number is 0.
Identifies a type of Model's prediction resources.
Used in:
Should not be used.
Resources that are dedicated to the [DeployedModel][google.cloud.aiplatform.v1.DeployedModel], and that need a higher degree of manual configuration.
Resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
Resources that can be shared by multiple [DeployedModels][google.cloud.aiplatform.v1.DeployedModel]. A pre-configured [DeploymentResourcePool][google.cloud.aiplatform.v1.DeploymentResourcePool] is required.
Represents export format supported by the Model. All formats export to Google Cloud Storage.
Used in:
Output only. The ID of the export format. The possible format IDs are: * `tflite` Used for Android mobile devices. * `edgetpu-tflite` Used for [Edge TPU](https://cloud.google.com/edge-tpu/) devices. * `tf-saved-model` A tensorflow model in SavedModel format. * `tf-js` A [TensorFlow.js](https://www.tensorflow.org/js) model that can be used in the browser and in Node.js using JavaScript. * `core-ml` Used for iOS mobile devices. * `custom-trained` A Model that was uploaded or trained by custom code.
Output only. The content of this Model that may be exported.
The Model content that can be exported.
Used in:
Should not be used.
Model artifact and any of its supported files. Will be exported to the location specified by the `artifactDestination` field of the [ExportModelRequest.output_config][google.cloud.aiplatform.v1.ExportModelRequest.output_config] object.
The container image that is to be used when deploying this Model. Will be exported to the location specified by the `imageDestination` field of the [ExportModelRequest.output_config][google.cloud.aiplatform.v1.ExportModelRequest.output_config] object.
Contains information about the original Model if this Model is a copy.
Used in:
Output only. The resource name of the Model this Model is a copy of, including the revision. Format: `projects/{project}/locations/{location}/models/{model_id}@{version_id}`
Specification of a container for serving predictions. Some fields in this message correspond to fields in the [Kubernetes Container v1 core specification](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
Used in:
, ,Required. Immutable. URI of the Docker image to be used as the custom container for serving predictions. This URI must identify an image in Artifact Registry or Container Registry. Learn more about the [container publishing requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#publishing), including permissions requirements for the Vertex AI Service Agent. The container image is ingested upon [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel], stored internally, and this original path is afterwards not used. To learn about the requirements for the Docker image itself, see [Custom container requirements](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#). You can use the URI to one of Vertex AI's [pre-built container images for prediction](https://cloud.google.com/vertex-ai/docs/predictions/pre-built-containers) in this field.
Immutable. Specifies the command that runs when the container starts. This overrides the container's [ENTRYPOINT](https://docs.docker.com/engine/reference/builder/#entrypoint). Specify this field as an array of executable and arguments, similar to a Docker `ENTRYPOINT`'s "exec" form, not its "shell" form. If you do not specify this field, then the container's `ENTRYPOINT` runs, in conjunction with the [args][google.cloud.aiplatform.v1.ModelContainerSpec.args] field or the container's [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd), if either exists. If this field is not specified and the container does not have an `ENTRYPOINT`, then refer to the Docker documentation about [how `CMD` and `ENTRYPOINT` interact](https://docs.docker.com/engine/reference/builder/#understand-how-cmd-and-entrypoint-interact). If you specify this field, then you can also specify the `args` field to provide additional arguments for this command. However, if you specify this field, then the container's `CMD` is ignored. See the [Kubernetes documentation about how the `command` and `args` fields interact with a container's `ENTRYPOINT` and `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#notes). In this field, you can reference [environment variables set by Vertex AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables) and environment variables set in the [env][google.cloud.aiplatform.v1.ModelContainerSpec.env] field. You cannot reference environment variables set in the Docker image. In order for environment variables to be expanded, reference them by using the following syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs from Bash variable expansion, which does not use parentheses. If a variable cannot be resolved, the reference in the input string is used unchanged. To avoid variable expansion, you can escape this syntax with `$$`; for example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds to the `command` field of the Kubernetes Containers [v1 core API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
Immutable. Specifies arguments for the command that runs when the container starts. This overrides the container's [`CMD`](https://docs.docker.com/engine/reference/builder/#cmd). Specify this field as an array of executable and arguments, similar to a Docker `CMD`'s "default parameters" form. If you don't specify this field but do specify the [command][google.cloud.aiplatform.v1.ModelContainerSpec.command] field, then the command from the `command` field runs without any additional arguments. See the [Kubernetes documentation about how the `command` and `args` fields interact with a container's `ENTRYPOINT` and `CMD`](https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#notes). If you don't specify this field and don't specify the `command` field, then the container's [`ENTRYPOINT`](https://docs.docker.com/engine/reference/builder/#cmd) and `CMD` determine what runs based on their default behavior. See the Docker documentation about [how `CMD` and `ENTRYPOINT` interact](https://docs.docker.com/engine/reference/builder/#understand-how-cmd-and-entrypoint-interact). In this field, you can reference [environment variables set by Vertex AI](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables) and environment variables set in the [env][google.cloud.aiplatform.v1.ModelContainerSpec.env] field. You cannot reference environment variables set in the Docker image. In order for environment variables to be expanded, reference them by using the following syntax: <code>$(<var>VARIABLE_NAME</var>)</code> Note that this differs from Bash variable expansion, which does not use parentheses. If a variable cannot be resolved, the reference in the input string is used unchanged. To avoid variable expansion, you can escape this syntax with `$$`; for example: <code>$$(<var>VARIABLE_NAME</var>)</code> This field corresponds to the `args` field of the Kubernetes Containers [v1 core API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
Immutable. List of environment variables to set in the container. After the container starts running, code running in the container can read these environment variables. Additionally, the [command][google.cloud.aiplatform.v1.ModelContainerSpec.command] and [args][google.cloud.aiplatform.v1.ModelContainerSpec.args] fields can reference these variables. Later entries in this list can also reference earlier entries. For example, the following example sets the variable `VAR_2` to have the value `foo bar`: ```json [ { "name": "VAR_1", "value": "foo" }, { "name": "VAR_2", "value": "$(VAR_1) bar" } ] ``` If you switch the order of the variables in the example, then the expansion does not occur. This field corresponds to the `env` field of the Kubernetes Containers [v1 core API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
Immutable. List of ports to expose from the container. Vertex AI sends any prediction requests that it receives to the first port on this list. Vertex AI also sends [liveness and health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#liveness) to this port. If you do not specify this field, it defaults to following value: ```json [ { "containerPort": 8080 } ] ``` Vertex AI does not use ports other than the first one listed. This field corresponds to the `ports` field of the Kubernetes Containers [v1 core API](https://kubernetes.io/docs/reference/generated/kubernetes-api/v1.23/#container-v1-core).
Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using [projects.locations.endpoints.predict][google.cloud.aiplatform.v1.PredictionService.Predict] to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s [ports][google.cloud.aiplatform.v1.ModelContainerSpec.ports] field. If you don't specify this field, it defaults to the following value when you [deploy this Model to an Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]: <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code> The placeholders in this value are replaced as follows: * <var>ENDPOINT</var>: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * <var>DEPLOYED_MODEL</var>: [DeployedModel.id][google.cloud.aiplatform.v1.DeployedModel.id] of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
Immutable. HTTP path on the container to send health checks to. Vertex AI intermittently sends GET requests to this path on the container's IP address and port to check that the container is healthy. Read more about [health checks](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#health). For example, if you set this field to `/bar`, then Vertex AI intermittently sends a GET request to the `/bar` path on the port of your container specified by the first value of this `ModelContainerSpec`'s [ports][google.cloud.aiplatform.v1.ModelContainerSpec.ports] field. If you don't specify this field, it defaults to the following value when you [deploy this Model to an Endpoint][google.cloud.aiplatform.v1.EndpointService.DeployModel]: <code>/v1/endpoints/<var>ENDPOINT</var>/deployedModels/<var>DEPLOYED_MODEL</var>:predict</code> The placeholders in this value are replaced as follows: * <var>ENDPOINT</var>: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * <var>DEPLOYED_MODEL</var>: [DeployedModel.id][google.cloud.aiplatform.v1.DeployedModel.id] of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).)
Immutable. List of ports to expose from the container. Vertex AI sends gRPC prediction requests that it receives to the first port on this list. Vertex AI also sends liveness and health checks to this port. If you do not specify this field, gRPC requests to the container will be disabled. Vertex AI does not use ports other than the first one listed. This field corresponds to the `ports` field of the Kubernetes Containers v1 core API.
Immutable. Deployment timeout. Limit for deployment timeout is 2 hours.
Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes.
Immutable. Specification for Kubernetes startup probe.
Immutable. Specification for Kubernetes readiness probe.
Immutable. Specification for Kubernetes liveness probe.
ModelDeploymentMonitoringBigQueryTable specifies the BigQuery table name as well as some information of the logs stored in this table.
Used in:
The source of log.
The type of log.
The created BigQuery table to store logs. Customer could do their own query & analysis. Format: `bq://<project_id>.model_deployment_monitoring_<endpoint_id>.<tolower(log_source)>_<tolower(log_type)>`
Output only. The schema version of the request/response logging BigQuery table. Default to v1 if unset.
Indicates where does the log come from.
Used in:
Unspecified source.
Logs coming from Training dataset.
Logs coming from Serving traffic.
Indicates what type of traffic does the log belong to.
Used in:
Unspecified type.
Predict logs.
Explain logs.
Represents a job that runs periodically to monitor the deployed models in an endpoint. It will analyze the logged training & prediction data to detect any abnormal behaviors.
Used as response type in: JobService.CreateModelDeploymentMonitoringJob, JobService.GetModelDeploymentMonitoringJob
Used as field type in:
, ,Output only. Resource name of a ModelDeploymentMonitoringJob.
Required. The user-defined name of the ModelDeploymentMonitoringJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. Display name of a ModelDeploymentMonitoringJob.
Required. Endpoint resource name. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
Output only. The detailed state of the monitoring job. When the job is still creating, the state will be 'PENDING'. Once the job is successfully created, the state will be 'RUNNING'. Pause the job, the state will be 'PAUSED'. Resume the job, the state will return to 'RUNNING'.
Output only. Schedule state when the monitoring job is in Running state.
Output only. Latest triggered monitoring pipeline metadata.
Required. The config for monitoring objectives. This is a per DeployedModel config. Each DeployedModel needs to be configured separately.
Required. Schedule config for running the monitoring job.
Required. Sample Strategy for logging.
Alert config for model monitoring.
YAML schema file uri describing the format of a single instance, which are given to format this Endpoint's prediction (and explanation). If not set, we will generate predict schema from collected predict requests.
Sample Predict instance, same format as [PredictRequest.instances][google.cloud.aiplatform.v1.PredictRequest.instances], this can be set as a replacement of [ModelDeploymentMonitoringJob.predict_instance_schema_uri][google.cloud.aiplatform.v1.ModelDeploymentMonitoringJob.predict_instance_schema_uri]. If not set, we will generate predict schema from collected predict requests.
YAML schema file uri describing the format of a single instance that you want Tensorflow Data Validation (TFDV) to analyze. If this field is empty, all the feature data types are inferred from [predict_instance_schema_uri][google.cloud.aiplatform.v1.ModelDeploymentMonitoringJob.predict_instance_schema_uri], meaning that TFDV will use the data in the exact format(data type) as prediction request/response. If there are any data type differences between predict instance and TFDV instance, this field can be used to override the schema. For models trained with Vertex AI, this field must be set as all the fields in predict instance formatted as string.
Output only. The created bigquery tables for the job under customer project. Customer could do their own query & analysis. There could be 4 log tables in maximum: 1. Training data logging predict request/response 2. Serving data logging predict request/response
The TTL of BigQuery tables in user projects which stores logs. A day is the basic unit of the TTL and we take the ceil of TTL/86400(a day). e.g. { second: 3600} indicates ttl = 1 day.
The labels with user-defined metadata to organize your ModelDeploymentMonitoringJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Output only. Timestamp when this ModelDeploymentMonitoringJob was created.
Output only. Timestamp when this ModelDeploymentMonitoringJob was updated most recently.
Output only. Timestamp when this monitoring pipeline will be scheduled to run for the next round.
Stats anomalies base folder path.
Customer-managed encryption key spec for a ModelDeploymentMonitoringJob. If set, this ModelDeploymentMonitoringJob and all sub-resources of this ModelDeploymentMonitoringJob will be secured by this key.
If true, the scheduled monitoring pipeline logs are sent to Google Cloud Logging, including pipeline status and anomalies detected. Please note the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging#pricing).
Output only. Only populated when the job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
Output only. Reserved for future use.
Output only. Reserved for future use.
All metadata of most recent monitoring pipelines.
Used in:
The time that most recent monitoring pipelines that is related to this run.
The status of the most recent monitoring pipeline.
The state to Specify the monitoring pipeline.
Used in:
Unspecified state.
The pipeline is picked up and wait to run.
The pipeline is offline and will be scheduled for next run.
The pipeline is running.
ModelDeploymentMonitoringObjectiveConfig contains the pair of deployed_model_id to ModelMonitoringObjectiveConfig.
Used in:
The DeployedModel ID of the objective config.
The objective config of for the modelmonitoring job of this deployed model.
The Model Monitoring Objective types.
Used in:
,Default value, should not be set.
Raw feature values' stats to detect skew between Training-Prediction datasets.
Raw feature values' stats to detect drift between Serving-Prediction datasets.
Feature attribution scores to detect skew between Training-Prediction datasets.
Feature attribution scores to detect skew between Prediction datasets collected within different time windows.
The config for scheduling monitoring job.
Used in:
Required. The model monitoring job scheduling interval. It will be rounded up to next full hour. This defines how often the monitoring jobs are triggered.
The time window of the prediction data being included in each prediction dataset. This window specifies how long the data should be collected from historical model results for each run. If not set, [ModelDeploymentMonitoringScheduleConfig.monitor_interval][google.cloud.aiplatform.v1.ModelDeploymentMonitoringScheduleConfig.monitor_interval] will be used. e.g. If currently the cutoff time is 2022-01-08 14:30:00 and the monitor_window is set to be 3600, then data from 2022-01-08 13:30:00 to 2022-01-08 14:30:00 will be retrieved and aggregated to calculate the monitoring statistics.
A collection of metrics calculated by comparing Model's predictions on all of the test data against annotations from the test data.
Used as response type in: ModelService.GetModelEvaluation, ModelService.ImportModelEvaluation
Used as field type in:
,Output only. The resource name of the ModelEvaluation.
The display name of the ModelEvaluation.
Points to a YAML file stored on Google Cloud Storage describing the [metrics][google.cloud.aiplatform.v1.ModelEvaluation.metrics] of this ModelEvaluation. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
Evaluation metrics of the Model. The schema of the metrics is stored in [metrics_schema_uri][google.cloud.aiplatform.v1.ModelEvaluation.metrics_schema_uri]
Output only. Timestamp when this ModelEvaluation was created.
All possible [dimensions][google.cloud.aiplatform.v1.ModelEvaluationSlice.Slice.dimension] of ModelEvaluationSlices. The dimensions can be used as the filter of the [ModelService.ListModelEvaluationSlices][google.cloud.aiplatform.v1.ModelService.ListModelEvaluationSlices] request, in the form of `slice.dimension = <dimension>`.
Points to a YAML file stored on Google Cloud Storage describing [EvaluatedDataItemView.data_item_payload][] and [EvaluatedAnnotation.data_item_payload][google.cloud.aiplatform.v1.EvaluatedAnnotation.data_item_payload]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). This field is not populated if there are neither EvaluatedDataItemViews nor EvaluatedAnnotations under this ModelEvaluation.
Points to a YAML file stored on Google Cloud Storage describing [EvaluatedDataItemView.predictions][], [EvaluatedDataItemView.ground_truths][], [EvaluatedAnnotation.predictions][google.cloud.aiplatform.v1.EvaluatedAnnotation.predictions], and [EvaluatedAnnotation.ground_truths][google.cloud.aiplatform.v1.EvaluatedAnnotation.ground_truths]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). This field is not populated if there are neither EvaluatedDataItemViews nor EvaluatedAnnotations under this ModelEvaluation.
Aggregated explanation metrics for the Model's prediction output over the data this ModelEvaluation uses. This field is populated only if the Model is evaluated with explanations, and only for AutoML tabular Models.
Describes the values of [ExplanationSpec][google.cloud.aiplatform.v1.ExplanationSpec] that are used for explaining the predicted values on the evaluated data.
The metadata of the ModelEvaluation. For the ModelEvaluation uploaded from Managed Pipeline, metadata contains a structured value with keys of "pipeline_job_id", "evaluation_dataset_type", "evaluation_dataset_path", "row_based_metrics_path".
Used in:
Explanation type. For AutoML Image Classification models, possible values are: * `image-integrated-gradients` * `image-xrai`
Explanation spec details.
A collection of metrics calculated by comparing Model's predictions on a slice of the test data against ground truth annotations.
Used as response type in: ModelService.GetModelEvaluationSlice
Used as field type in:
,Output only. The resource name of the ModelEvaluationSlice.
Output only. The slice of the test data that is used to evaluate the Model.
Output only. Points to a YAML file stored on Google Cloud Storage describing the [metrics][google.cloud.aiplatform.v1.ModelEvaluationSlice.metrics] of this ModelEvaluationSlice. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject).
Output only. Sliced evaluation metrics of the Model. The schema of the metrics is stored in [metrics_schema_uri][google.cloud.aiplatform.v1.ModelEvaluationSlice.metrics_schema_uri]
Output only. Timestamp when this ModelEvaluationSlice was created.
Output only. Aggregated explanation metrics for the Model's prediction output over the data this ModelEvaluation uses. This field is populated only if the Model is evaluated with explanations, and only for tabular Models.
Definition of a slice.
Used in:
Output only. The dimension of the slice. Well-known dimensions are: * `annotationSpec`: This slice is on the test data that has either ground truth or prediction with [AnnotationSpec.display_name][google.cloud.aiplatform.v1.AnnotationSpec.display_name] equals to [value][google.cloud.aiplatform.v1.ModelEvaluationSlice.Slice.value]. * `slice`: This slice is a user customized slice defined by its SliceSpec.
Output only. The value of the dimension in this slice.
Output only. Specification for how the data was sliced.
Specification for how the data should be sliced.
Used in:
Mapping configuration for this SliceSpec. The key is the name of the feature. By default, the key will be prefixed by "instance" as a dictionary prefix for Vertex Batch Predictions output format.
A range of values for slice(s). `low` is inclusive, `high` is exclusive.
Used in:
Inclusive low value for the range.
Exclusive high value for the range.
Specification message containing the config for this SliceSpec. When `kind` is selected as `value` and/or `range`, only a single slice will be computed. When `all_values` is present, a separate slice will be computed for each possible label/value for the corresponding key in `config`. Examples, with feature zip_code with values 12345, 23334, 88888 and feature country with values "US", "Canada", "Mexico" in the dataset: Example 1: { "zip_code": { "value": { "float_value": 12345.0 } } } A single slice for any data with zip_code 12345 in the dataset. Example 2: { "zip_code": { "range": { "low": 12345, "high": 20000 } } } A single slice containing data where the zip_codes between 12345 and 20000 For this example, data with the zip_code of 12345 will be in this slice. Example 3: { "zip_code": { "range": { "low": 10000, "high": 20000 } }, "country": { "value": { "string_value": "US" } } } A single slice containing data where the zip_codes between 10000 and 20000 has the country "US". For this example, data with the zip_code of 12345 and country "US" will be in this slice. Example 4: { "country": {"all_values": { "value": true } } } Three slices are computed, one for each unique country in the dataset. Example 5: { "country": { "all_values": { "value": true } }, "zip_code": { "value": { "float_value": 12345.0 } } } Three slices are computed, one for each unique country in the dataset where the zip_code is also 12345. For this example, data with zip_code 12345 and country "US" will be in one slice, zip_code 12345 and country "Canada" in another slice, and zip_code 12345 and country "Mexico" in another slice, totaling 3 slices.
Used in:
A unique specific value for a given feature. Example: `{ "value": { "string_value": "12345" } }`
A range of values for a numerical feature. Example: `{"range":{"low":10000.0,"high":50000.0}}` will capture 12345 and 23334 in the slice.
If all_values is set to true, then all possible labels of the keyed feature will have another slice computed. Example: `{"all_values":{"value":true}}`
Single value that supports strings and floats.
Used in:
String type.
Float type.
Aggregated explanation metrics for a Model over a set of instances.
Used in:
,Output only. Aggregated attributions explaining the Model's prediction outputs over the set of instances. The attributions are grouped by outputs. For Models that predict only one output, such as regression Models that predict only one score, there is only one attibution that explains the predicted output. For Models that predict multiple outputs, such as multiclass Models that predict multiple classes, each element explains one specific item. [Attribution.output_index][google.cloud.aiplatform.v1.Attribution.output_index] can be used to identify which output this attribution is explaining. The [baselineOutputValue][google.cloud.aiplatform.v1.Attribution.baseline_output_value], [instanceOutputValue][google.cloud.aiplatform.v1.Attribution.instance_output_value] and [featureAttributions][google.cloud.aiplatform.v1.Attribution.feature_attributions] fields are averaged over the test data. NOTE: Currently AutoML tabular classification Models produce only one attribution, which averages attributions over all the classes it predicts. [Attribution.approximation_error][google.cloud.aiplatform.v1.Attribution.approximation_error] is not populated.
Contains information about the source of the models generated from Model Garden.
Used in:
Required. The model garden source model resource name.
Optional. The model garden source model version ID.
Optional. Whether to avoid pulling the model from the HF cache.
The alert config for model monitoring.
Used in:
Email alert config.
Dump the anomalies to Cloud Logging. The anomalies will be put to json payload encoded from proto [ModelMonitoringStatsAnomalies][google.cloud.aiplatform.v1.ModelMonitoringStatsAnomalies]. This can be further synced to Pub/Sub or any other services supported by Cloud Logging.
Resource names of the NotificationChannels to send alert. Must be of the format `projects/<project_id_or_number>/notificationChannels/<channel_id>`
The config for email alert.
Used in:
The email addresses to send the alert.
The objective configuration for model monitoring, including the information needed to detect anomalies for one particular model.
Used in:
Training dataset for models. This field has to be set only if TrainingPredictionSkewDetectionConfig is specified.
The config for skew between training data and prediction data.
The config for drift of prediction data.
The config for integrating with Vertex Explainable AI.
The config for integrating with Vertex Explainable AI. Only applicable if the Model has explanation_spec populated.
Used in:
If want to analyze the Vertex Explainable AI feature attribute scores or not. If set to true, Vertex AI will log the feature attributions from explain response and do the skew/drift detection for them.
Predictions generated by the BatchPredictionJob using baseline dataset.
Output from [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob] for Model Monitoring baseline dataset, which can be used to generate baseline attribution scores.
Used in:
The configuration specifying of BatchExplain job output. This can be used to generate the baseline of feature attribution scores.
Cloud Storage location for BatchExplain output.
BigQuery location for BatchExplain output.
The storage format of the predictions generated BatchPrediction job.
The storage format of the predictions generated BatchPrediction job.
Used in:
Should not be set.
Predictions are in JSONL files.
Predictions are in BigQuery.
The config for Prediction data drift detection.
Used in:
Key is the feature name and value is the threshold. If a feature needs to be monitored for drift, a value threshold must be configured for that feature. The threshold here is against feature distribution distance between different time windws.
Key is the feature name and value is the threshold. The threshold here is against attribution score distance between different time windows.
Drift anomaly detection threshold used by all features. When the per-feature thresholds are not set, this field can be used to specify a threshold for all features.
Training Dataset information.
Used in:
The resource name of the Dataset used to train this Model.
The Google Cloud Storage uri of the unmanaged Dataset used to train this Model.
The BigQuery table of the unmanaged Dataset used to train this Model.
Data format of the dataset, only applicable if the input is from Google Cloud Storage. The possible formats are: "tf-record" The source file is a TFRecord file. "csv" The source file is a CSV file. "jsonl" The source file is a JSONL file.
The target field name the model is to predict. This field will be excluded when doing Predict and (or) Explain for the training data.
Strategy to sample data from Training Dataset. If not set, we process the whole dataset.
The config for Training & Prediction data skew detection. It specifies the training dataset sources and the skew detection parameters.
Used in:
Key is the feature name and value is the threshold. If a feature needs to be monitored for skew, a value threshold must be configured for that feature. The threshold here is against feature distribution distance between the training and prediction feature.
Key is the feature name and value is the threshold. The threshold here is against attribution score distance between the training and prediction feature.
Skew anomaly detection threshold used by all features. When the per-feature thresholds are not set, this field can be used to specify a threshold for all features.
Statistics and anomalies generated by Model Monitoring.
Used in:
Model Monitoring Objective those stats and anomalies belonging to.
Deployed Model ID.
Number of anomalies within all stats.
A list of historical Stats and Anomalies generated for all Features.
Historical Stats (and Anomalies) for a specific Feature.
Used in:
Display Name of the Feature.
Threshold for anomaly detection.
Stats calculated for the Training Dataset.
A list of historical stats generated by different time window's Prediction Dataset.
Detail description of the source information of the model.
Used in:
Type of the model source.
If this Model is copy of another Model. If true then [source_type][google.cloud.aiplatform.v1.ModelSourceInfo.source_type] pertains to the original.
Source of the model. Different from `objective` field, this `ModelSourceType` enum indicates the source from which the model was accessed or obtained, whereas the `objective` indicates the overall aim or function of this model.
Used in:
Should not be used.
The Model is uploaded by automl training pipeline.
The Model is uploaded by user or custom training pipeline.
The Model is registered and sync'ed from BigQuery ML.
The Model is saved or tuned from Model Garden.
The Model is saved or tuned from Genie.
The Model is uploaded by text embedding finetuning pipeline.
The Model is saved or tuned from Marketplace.
A proto representation of a Spanner-stored ModelVersionCheckpoint. The meaning of the fields is equivalent to their in-Spanner counterparts.
Used in:
The ID of the checkpoint.
The epoch of the checkpoint.
The step of the checkpoint.
Runtime operation information for [IndexEndpointService.MutateDeployedIndex][google.cloud.aiplatform.v1.IndexEndpointService.MutateDeployedIndex].
The operation generic information.
The unique index id specified by user
Response message for [IndexEndpointService.MutateDeployedIndex][google.cloud.aiplatform.v1.IndexEndpointService.MutateDeployedIndex].
The DeployedIndex that had been updated in the IndexEndpoint.
Runtime operation information for [EndpointService.MutateDeployedModel][google.cloud.aiplatform.v1.EndpointService.MutateDeployedModel].
The operation generic information.
Response message for [EndpointService.MutateDeployedModel][google.cloud.aiplatform.v1.EndpointService.MutateDeployedModel].
The DeployedModel that's being mutated.
Represents a Neural Architecture Search (NAS) job.
Used as response type in: JobService.CreateNasJob, JobService.GetNasJob
Used as field type in:
,Output only. Resource name of the NasJob.
Required. The display name of the NasJob. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Required. The specification of a NasJob.
Output only. Output of the NasJob.
Output only. The detailed state of the job.
Output only. Time when the NasJob was created.
Output only. Time when the NasJob for the first time entered the `JOB_STATE_RUNNING` state.
Output only. Time when the NasJob entered any of the following states: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`.
Output only. Time when the NasJob was most recently updated.
Output only. Only populated when job's state is JOB_STATE_FAILED or JOB_STATE_CANCELLED.
The labels with user-defined metadata to organize NasJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Customer-managed encryption key options for a NasJob. If this is set, then all resources created by the NasJob will be encrypted with the provided encryption key.
Optional. Enable a separation of Custom model training and restricted image training for tenant project.
Output only. Reserved for future use.
Output only. Reserved for future use.
Represents a uCAIP NasJob output.
Used in:
The output of this Neural Architecture Search (NAS) job.
Output only. The output of this multi-trial Neural Architecture Search (NAS) job.
The output of a multi-trial Neural Architecture Search (NAS) jobs.
Used in:
Output only. List of NasTrials that were started as part of search stage.
Output only. List of NasTrials that were started as part of train stage.
Represents the spec of a NasJob.
Used in:
The Neural Architecture Search (NAS) algorithm specification.
The spec of multi-trial algorithms.
The ID of the existing NasJob in the same Project and Location which will be used to resume search. search_space_spec and nas_algorithm_spec are obtained from previous NasJob hence should not provide them again for this NasJob.
It defines the search space for Neural Architecture Search (NAS).
The spec of multi-trial Neural Architecture Search (NAS).
Used in:
The multi-trial Neural Architecture Search (NAS) algorithm type. Defaults to `REINFORCEMENT_LEARNING`.
Metric specs for the NAS job. Validation for this field is done at `multi_trial_algorithm_spec` field.
Required. Spec for search trials.
Spec for train trials. Top N [TrainTrialSpec.max_parallel_trial_count] search trials will be trained for every M [TrainTrialSpec.frequency] trials searched.
Represents a metric to optimize.
Used in:
Required. The ID of the metric. Must not contain whitespaces.
Required. The optimization goal of the metric.
The available types of optimization goals.
Used in:
Goal Type will default to maximize.
Maximize the goal metric.
Minimize the goal metric.
The available types of multi-trial algorithms.
Used in:
Defaults to `REINFORCEMENT_LEARNING`.
The Reinforcement Learning Algorithm for Multi-trial Neural Architecture Search (NAS).
The Grid Search Algorithm for Multi-trial Neural Architecture Search (NAS).
Represent spec for search trials.
Used in:
Required. The spec of a search trial job. The same spec applies to all search trials.
Required. The maximum number of Neural Architecture Search (NAS) trials to run.
Required. The maximum number of trials to run in parallel.
The number of failed trials that need to be seen before failing the NasJob. If set to 0, Vertex AI decides how many trials must fail before the whole job fails.
Represent spec for train trials.
Used in:
Required. The spec of a train trial job. The same spec applies to all train trials.
Required. The maximum number of trials to run in parallel.
Required. Frequency of search trials to start train stage. Top N [TrainTrialSpec.max_parallel_trial_count] search trials will be trained for every M [TrainTrialSpec.frequency] trials searched.
Represents a uCAIP NasJob trial.
Used in:
,Output only. The identifier of the NasTrial assigned by the service.
Output only. The detailed state of the NasTrial.
Output only. The final measurement containing the objective value.
Output only. Time when the NasTrial was started.
Output only. Time when the NasTrial's status changed to `SUCCEEDED` or `INFEASIBLE`.
Describes a NasTrial state.
Used in:
The NasTrial state is unspecified.
Indicates that a specific NasTrial has been requested, but it has not yet been suggested by the service.
Indicates that the NasTrial has been suggested.
Indicates that the NasTrial should stop according to the service.
Indicates that the NasTrial is completed successfully.
Indicates that the NasTrial should not be attempted again. The service will set a NasTrial to INFEASIBLE when it's done but missing the final_measurement.
Represents a NasTrial details along with its parameters. If there is a corresponding train NasTrial, the train NasTrial is also returned.
Used as response type in: JobService.GetNasTrialDetail
Used as field type in:
Output only. Resource name of the NasTrialDetail.
The parameters for the NasJob NasTrial.
The requested search NasTrial.
The train NasTrial corresponding to [search_trial][google.cloud.aiplatform.v1.NasTrialDetail.search_trial]. Only populated if [search_trial][google.cloud.aiplatform.v1.NasTrialDetail.search_trial] is used for training.
A query to find a number of similar entities.
Used in:
Optional. The entity id whose similar entities should be searched for. If embedding is set, search will use embedding instead of entity_id.
Optional. The embedding vector that be used for similar search.
Optional. The number of similar entities to be retrieved from feature view for each query.
Optional. The list of string filters.
Optional. The list of numeric filters.
Optional. Crowding is a constraint on a neighbor list produced by nearest neighbor search requiring that no more than sper_crowding_attribute_neighbor_count of the k neighbors returned have the same value of crowding_attribute. It's used for improving result diversity.
Optional. Parameters that can be set to tune query on the fly.
The embedding vector.
Used in:
Optional. Individual value in the embedding.
Numeric filter is used to search a subset of the entities by using boolean rules on numeric columns. For example: Database Point 0: {name: "a" value_int: 42} {name: "b" value_float: 1.0} Database Point 1: {name: "a" value_int: 10} {name: "b" value_float: 2.0} Database Point 2: {name: "a" value_int: -1} {name: "b" value_float: 3.0} Query: {name: "a" value_int: 12 operator: LESS} // Matches Point 1, 2 {name: "b" value_float: 2.0 operator: EQUAL} // Matches Point 1
Used in:
The type of Value must be consistent for all datapoints with a given name. This is verified at runtime.
int value type.
float value type.
double value type.
Required. Column name in BigQuery that used as filters.
Optional. This MUST be specified for queries and must NOT be specified for database points.
Datapoints for which Operator is true relative to the query's Value field will be allowlisted.
Used in:
Unspecified operator.
Entities are eligible if their value is < the query's.
Entities are eligible if their value is <= the query's.
Entities are eligible if their value is == the query's.
Entities are eligible if their value is >= the query's.
Entities are eligible if their value is > the query's.
Entities are eligible if their value is != the query's.
Parameters that can be overrided in each query to tune query latency and recall.
Used in:
Optional. The number of neighbors to find via approximate search before exact reordering is performed; if set, this value must be > neighbor_count.
Optional. The fraction of the number of leaves to search, set at query time allows user to tune search performance. This value increase result in both search accuracy and latency increase. The value should be between 0.0 and 1.0.
String filter is used to search a subset of the entities by using boolean rules on string columns. For example: if a query specifies string filter with 'name = color, allow_tokens = {red, blue}, deny_tokens = {purple}',' then that query will match entities that are red or blue, but if those points are also purple, then they will be excluded even if they are red/blue. Only string filter is supported for now, numeric filter will be supported in the near future.
Used in:
Required. Column names in BigQuery that used as filters.
Optional. The allowed tokens.
Optional. The denied tokens.
Runtime operation metadata with regard to Matching Engine Index.
Used in:
,The validation stats of the content (per file) to be inserted or updated on the Matching Engine Index resource. Populated if contentsDeltaUri is provided as part of [Index.metadata][google.cloud.aiplatform.v1.Index.metadata]. Please note that, currently for those files that are broken or has unsupported file format, we will not have the stats for those files.
The ingested data size in bytes.
Used in:
Cloud Storage URI pointing to the original file in user's bucket.
Number of records in this file that were successfully processed.
Number of records in this file we skipped due to validate errors.
The detail information of the partial failures encountered for those invalid records that couldn't be parsed. Up to 50 partial errors will be reported.
Number of sparse records in this file that were successfully processed.
Number of sparse records in this file we skipped due to validate errors.
Used in:
The error type of this record.
A human-readable message that is shown to the user to help them fix the error. Note that this message may change from time to time, your code should check against error_type as the source of truth.
Cloud Storage URI pointing to the original file in user's bucket.
Empty if the embedding id is failed to parse.
The original content of this record.
Used in:
Default, shall not be used.
The record is empty.
Invalid json format.
Invalid csv format.
Invalid avro format.
The embedding id is not valid.
The size of the dense embedding vectors does not match with the specified dimension.
The `namespace` field is missing.
Generic catch-all error. Only used for validation failure where the root cause cannot be easily retrieved programmatically.
There are multiple restricts with the same `namespace` value.
Numeric restrict has operator specified in datapoint.
Numeric restrict has multiple values specified.
Numeric restrict has invalid numeric value specified.
File is not in UTF_8 format.
Error parsing sparse dimensions field.
Token restrict value is invalid.
Invalid sparse embedding.
Invalid dense embedding.
Nearest neighbors for one query.
Used in:
All its neighbors.
A neighbor of the query vector.
Used in:
The id of the similar entity.
The distance between the neighbor and the query vector.
The attributes of the neighbor, e.g. filters, crowding and metadata Note that full entities are returned only when "return_full_entity" is set to true. Otherwise, only the "entity_id" and "distance" fields are populated.
Neighbors for example-based explanations.
Used in:
Output only. The neighbor id.
Output only. The neighbor distance.
Network spec.
Used in:
, ,Whether to enable public internet access. Default false.
The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks)
The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}`
Represents a mount configuration for Network File System (NFS) to mount.
Used in:
Required. IP address of the NFS server.
Required. Source path exported from NFS server. Has to start with '/', and combined with the ip address, it indicates the source mount path in the form of `server:path`
Required. Destination mount path. The NFS will be mounted for the user under /mnt/nfs/<mount_point>
The euc configuration of NotebookRuntimeTemplate.
Used in:
,Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate.
Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA.
NotebookExecutionJob represents an instance of a notebook execution.
Used as response type in: NotebookService.GetNotebookExecutionJob
Used as field type in:
,The input notebook.
The Dataform Repository pointing to a single file notebook repository.
The Cloud Storage url pointing to the ipynb file. Format: `gs://bucket/notebook_file.ipynb`
The contents of an input notebook file.
The compute config to use for an execution job.
The NotebookRuntimeTemplate to source compute configuration from.
The custom compute configuration for an execution job.
The location to store the notebook execution result.
The Cloud Storage location to upload the result to. Format: `gs://bucket-name`
The identity to run the execution as.
The user email to run the execution as. Only supported by Colab runtimes.
The service account to run the execution as.
Runtime environment for the notebook execution job. If unspecified, the default runtime of Colab is used.
The Workbench runtime configuration to use for the notebook execution.
Output only. The resource name of this NotebookExecutionJob. Format: `projects/{project_id}/locations/{location}/notebookExecutionJobs/{job_id}`
The display name of the NotebookExecutionJob. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Max running time of the execution job in seconds (default 86400s / 24 hrs).
Output only. The Schedule resource name if this job is triggered by one. Format: `projects/{project_id}/locations/{location}/schedules/{schedule_id}`
Output only. The state of the NotebookExecutionJob.
Output only. Populated when the NotebookExecutionJob is completed. When there is an error during notebook execution, the error details are populated.
Output only. Timestamp when this NotebookExecutionJob was created.
Output only. Timestamp when this NotebookExecutionJob was most recently updated.
The labels with user-defined metadata to organize NotebookExecutionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
The name of the kernel to use during notebook execution. If unset, the default kernel is used.
Customer-managed encryption key spec for the notebook execution job. This field is auto-populated if the [NotebookRuntimeTemplate][google.cloud.aiplatform.v1.NotebookRuntimeTemplate] has an encryption spec.
Compute configuration to use for an execution job.
Used in:
The specification of a single machine for the execution job.
The specification of a persistent disk to attach for the execution job.
The network configuration to use for the execution job.
The Dataform Repository containing the input notebook.
Used in:
The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}`
The commit SHA to read repository with. If unset, the file will be read at HEAD.
The content of the input notebook in ipynb format.
Used in:
The base64-encoded contents of the input notebook file.
The Cloud Storage uri for the input notebook.
Used in:
The Cloud Storage uri pointing to the ipynb file. Format: `gs://bucket/notebook_file.ipynb`
The version of the Cloud Storage object to read. If unset, the current version of the object is read. See https://cloud.google.com/storage/docs/metadata#generation-number.
Configuration for a Workbench Instances-based environment.
Used in:
(message has no fields)
Views for Get/List NotebookExecutionJob
Used in:
,When unspecified, the API defaults to the BASIC view.
Includes all fields except for direct notebook inputs.
Includes all fields.
The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field.
Used in:
,Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60.
Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate.
A runtime is a virtual machine allocated to a particular user for a particular Notebook file on temporary basis with lifetime limited to 24 hours.
Used as response type in: NotebookService.GetNotebookRuntime
Used as field type in:
,Output only. The resource name of the NotebookRuntime.
Required. The user email of the NotebookRuntime.
Output only. The pointer to NotebookRuntimeTemplate this NotebookRuntime is created from.
Output only. The proxy endpoint used to access the NotebookRuntime.
Output only. Timestamp when this NotebookRuntime was created.
Output only. Timestamp when this NotebookRuntime was most recently updated.
Output only. The health state of the NotebookRuntime.
Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the NotebookRuntime.
Output only. Deprecated: This field is no longer used and the "Vertex AI Notebook Service Account" (service-PROJECT_NUMBER@gcp-sa-aiplatform-vm.iam.gserviceaccount.com) is used for the runtime workload identity. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-notebook-service-account for more details. The service account that the NotebookRuntime workload runs as.
Output only. The runtime (instance) state of the NotebookRuntime.
Output only. Whether NotebookRuntime is upgradable.
The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex.
Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade.
Output only. The VM os image version of NotebookRuntime.
Output only. The type of the notebook runtime.
Output only. The specification of a single machine used by the notebook runtime.
Output only. The specification of [persistent disk][https://cloud.google.com/compute/docs/disks/persistent-disks] attached to the notebook runtime as data disk storage.
Output only. Network spec of the notebook runtime.
Output only. The idle shutdown configuration of the notebook runtime.
Output only. EUC configuration of the notebook runtime.
Output only. Runtime Shielded VM spec.
Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)).
Output only. Software config of the notebook runtime.
Output only. Customer-managed encryption key spec for the notebook runtime.
Output only. Reserved for future use.
Output only. Reserved for future use.
The substate of the NotebookRuntime to display health information.
Used in:
Unspecified health state.
NotebookRuntime is in healthy state. Applies to ACTIVE state.
NotebookRuntime is in unhealthy state. Applies to ACTIVE state.
The substate of the NotebookRuntime to display state of runtime. The resource of NotebookRuntime is in ACTIVE state for these sub state.
Used in:
Unspecified runtime state.
NotebookRuntime is in running state.
NotebookRuntime is in starting state.
NotebookRuntime is in stopping state.
NotebookRuntime is in stopped state.
NotebookRuntime is in upgrading state. It is in the middle of upgrading process.
NotebookRuntime was unable to start/stop properly.
NotebookRuntime is in invalid state. Cannot be recovered.
A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template.
Used as response type in: NotebookService.GetNotebookRuntimeTemplate, NotebookService.UpdateNotebookRuntimeTemplate
Used as field type in:
, ,The resource name of the NotebookRuntimeTemplate.
Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters.
The description of the NotebookRuntimeTemplate.
Output only. Deprecated: This field has no behavior. Use notebook_runtime_type = 'ONE_CLICK' instead. The default template to use if not specified.
Optional. Immutable. The specification of a single machine for the template.
Optional. The specification of [persistent disk][https://cloud.google.com/compute/docs/disks/persistent-disks] attached to the runtime as data disk storage.
Optional. Network spec.
Deprecated: This field is ignored and the "Vertex AI Notebook Service Account" (service-PROJECT_NUMBER@gcp-sa-aiplatform-vm.iam.gserviceaccount.com) is used for the runtime workload identity. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-notebook-service-account for more details. For NotebookExecutionJob, use NotebookExecutionJob.service_account instead. The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled.
EUC configuration of the NotebookRuntimeTemplate.
Output only. Timestamp when this NotebookRuntimeTemplate was created.
Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated.
Optional. Immutable. The type of the notebook runtime template.
Optional. Immutable. Runtime Shielded VM spec.
Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)).
Customer-managed encryption key spec for the notebook runtime.
Optional. The notebook software configuration of the notebook runtime.
Points to a NotebookRuntimeTemplateRef.
Used in:
Immutable. A resource name of the NotebookRuntimeTemplate.
Represents a notebook runtime type.
Used in:
,Unspecified notebook runtime type, NotebookRuntimeType will default to USER_DEFINED.
runtime or template with coustomized configurations from user.
runtime or template with system defined configurations.
Notebook Software Config.
Used in:
,Optional. Environment variables to be passed to the container. Maximum limit is 100.
Optional. Post startup script config.
PSC config that is used to automatically create forwarding rule via ServiceConnectionMap.
Used in:
Required. Project id used to create forwarding rule.
Required. The full name of the Google Compute Engine [network](https://cloud.google.com/compute/docs/networks-and-firewalls#networks). [Format](https://cloud.google.com/compute/docs/reference/rest/v1/networks/insert): `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in '12345', and {network} is network name.
Pairwise prediction autorater preference.
Used in:
, ,Unspecified prediction choice.
Baseline prediction wins
Candidate prediction wins
Winner cannot be determined
Input for pairwise metric.
Used in:
Required. Spec for pairwise metric.
Required. Pairwise metric instance.
Pairwise metric instance. Usually one instance corresponds to one row in an evaluation dataset.
Used in:
Instance for pairwise metric.
Instance specified as a json string. String key-value pairs are expected in the json_instance to render PairwiseMetricSpec.instance_prompt_template.
Spec for pairwise metric result.
Used in:
Output only. Pairwise metric choice.
Output only. Explanation for pairwise metric score.
Spec for pairwise metric.
Used in:
Required. Metric prompt template for pairwise metric.
Input for pairwise question answering quality metric.
Used in:
Required. Spec for pairwise question answering quality score metric.
Required. Pairwise question answering quality instance.
Spec for pairwise question answering quality instance.
Used in:
Required. Output of the candidate model.
Required. Output of the baseline model.
Optional. Ground truth used to compare against the prediction.
Required. Text to answer the question.
Required. Question Answering prompt for LLM.
Spec for pairwise question answering quality result.
Used in:
Output only. Pairwise question answering prediction choice.
Output only. Explanation for question answering quality score.
Output only. Confidence for question answering quality score.
Spec for pairwise question answering quality score metric.
Used in:
Optional. Whether to use instance.reference to compute question answering quality.
Optional. Which version to use for evaluation.
Input for pairwise summarization quality metric.
Used in:
Required. Spec for pairwise summarization quality score metric.
Required. Pairwise summarization quality instance.
Spec for pairwise summarization quality instance.
Used in:
Required. Output of the candidate model.
Required. Output of the baseline model.
Optional. Ground truth used to compare against the prediction.
Required. Text to be summarized.
Required. Summarization prompt for LLM.
Spec for pairwise summarization quality result.
Used in:
Output only. Pairwise summarization prediction choice.
Output only. Explanation for summarization quality score.
Output only. Confidence for summarization quality score.
Spec for pairwise summarization quality score metric.
Used in:
Optional. Whether to use instance.reference to compute pairwise summarization quality.
Optional. Which version to use for evaluation.
A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes.
Used in:
Optional. Text part (can be code).
Optional. Inlined bytes data.
Optional. URI based data.
Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values.
Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model.
Optional. Code generated by the model that is meant to be executed.
Optional. Result of executing the [ExecutableCode].
Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data.
Represents the spec of [persistent disk][https://cloud.google.com/compute/docs/disks/persistent-disks] options.
Used in:
, ,Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk)
Size in GB of the disk (default is 100GB).
Represents long-lasting resources that are dedicated to users to runs custom workloads. A PersistentResource can have multiple node pools and each node pool can have its own machine spec.
Used as response type in: PersistentResourceService.GetPersistentResource
Used as field type in:
, ,Immutable. Resource name of a PersistentResource.
Optional. The display name of the PersistentResource. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Required. The spec of the pools of different resources.
Output only. The detailed state of a Study.
Output only. Only populated when persistent resource's state is `STOPPING` or `ERROR`.
Output only. Time when the PersistentResource was created.
Output only. Time when the PersistentResource for the first time entered the `RUNNING` state.
Output only. Time when the PersistentResource was most recently updated.
Optional. The labels with user-defined metadata to organize PersistentResource. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Optional. The full name of the Compute Engine [network](/compute/docs/networks-and-firewalls#networks) to peered with Vertex AI to host the persistent resources. For example, `projects/12345/global/networks/myVPC`. [Format](/compute/docs/reference/rest/v1/networks/insert) is of the form `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in `12345`, and {network} is a network name. To specify this field, you must have already [configured VPC Network Peering for Vertex AI](https://cloud.google.com/vertex-ai/docs/general/vpc-peering). If this field is left unspecified, the resources aren't peered with any network.
Optional. Customer-managed encryption key spec for a PersistentResource. If set, this PersistentResource and all sub-resources of this PersistentResource will be secured by this key.
Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration.
Output only. Runtime information of the Persistent Resource.
Optional. A list of names for the reserved IP ranges under the VPC network that can be used for this persistent resource. If set, we will deploy the persistent resource within the provided IP ranges. Otherwise, the persistent resource is deployed to any IP ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
Describes the PersistentResource state.
Used in:
Not set.
The PROVISIONING state indicates the persistent resources is being created.
The RUNNING state indicates the persistent resource is healthy and fully usable.
The STOPPING state indicates the persistent resource is being deleted.
The ERROR state indicates the persistent resource may be unusable. Details can be found in the `error` field.
The REBOOTING state indicates the persistent resource is being rebooted (PR is not available right now but is expected to be ready again later).
The UPDATING state indicates the persistent resource is being updated.
Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
Used in:
Default value, and follows fail slow behavior.
Indicates that the pipeline should continue to run until all possible tasks have been scheduled and completed.
Indicates that the pipeline should stop scheduling new tasks after a task has failed.
An instance of a machine learning PipelineJob.
Used as response type in: PipelineService.CreatePipelineJob, PipelineService.GetPipelineJob
Used as field type in:
, , ,Output only. The resource name of the PipelineJob.
The display name of the Pipeline. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Output only. Pipeline creation time.
Output only. Pipeline start time.
Output only. Pipeline end time.
Output only. Timestamp when this PipelineJob was most recently updated.
The spec of the pipeline.
Output only. The detailed state of the job.
Output only. The details of pipeline run. Not available in the list view.
Output only. The error that occurred during pipeline execution. Only populated when the pipeline's state is FAILED or CANCELLED.
The labels with user-defined metadata to organize PipelineJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. Note there is some reserved label key for Vertex AI Pipelines. - `vertex-ai-pipelines-run-billing-id`, user set value will get overrided.
Runtime config of the pipeline.
Customer-managed encryption key spec for a pipelineJob. If set, this PipelineJob and all of its sub-resources will be secured by this key.
The service account that the pipeline workload runs as. If not specified, the Compute Engine default service account in the project will be used. See https://cloud.google.com/compute/docs/access/service-accounts#default_service_account Users starting the pipeline must have the `iam.serviceAccounts.actAs` permission on this service account.
The full name of the Compute Engine [network](/compute/docs/networks-and-firewalls#networks) to which the Pipeline Job's workload should be peered. For example, `projects/12345/global/networks/myVPC`. [Format](/compute/docs/reference/rest/v1/networks/insert) is of the form `projects/{project}/global/networks/{network}`. Where {project} is a project number, as in `12345`, and {network} is a network name. Private services access must already be configured for the network. Pipeline job will apply the network configuration to the Google Cloud resources being launched, if applied, such as Vertex AI Training or Dataflow job. If left unspecified, the workload is not peered with any network.
A list of names for the reserved ip ranges under the VPC network that can be used for this Pipeline Job's workload. If set, we will deploy the Pipeline Job's workload within the provided ip ranges. Otherwise, the job will be deployed to any ip ranges under the provided VPC network. Example: ['vertex-ai-ip-range'].
A template uri from where the [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec], if empty, will be downloaded. Currently, only uri from Vertex Template Registry & Gallery is supported. Reference to https://cloud.google.com/vertex-ai/docs/pipelines/create-pipeline-template.
Output only. Pipeline template metadata. Will fill up fields if [PipelineJob.template_uri][google.cloud.aiplatform.v1.PipelineJob.template_uri] is from supported template registry.
Output only. The schedule resource name. Only returned if the Pipeline is created by Schedule API.
Optional. Whether to do component level validations before job creation.
The runtime config of a PipelineJob.
Used in:
Deprecated. Use [RuntimeConfig.parameter_values][google.cloud.aiplatform.v1.PipelineJob.RuntimeConfig.parameter_values] instead. The runtime parameters of the PipelineJob. The parameters will be passed into [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] to replace the placeholders at runtime. This field is used by pipelines built using `PipelineJob.pipeline_spec.schema_version` 2.0.0 or lower, such as pipelines built using Kubeflow Pipelines SDK 1.8 or lower.
Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the pipeline. It is used by the system to generate the paths of output artifacts. The artifact paths are generated with a sub-path pattern `{job_id}/{task_id}/{output_key}` under the specified output directory. The service account specified in this pipeline must have the `storage.objects.get` and `storage.objects.create` permissions for this bucket.
The runtime parameters of the PipelineJob. The parameters will be passed into [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] to replace the placeholders at runtime. This field is used by pipelines built using `PipelineJob.pipeline_spec.schema_version` 2.1.0, such as pipelines built using Kubeflow Pipelines SDK 1.9 or higher and the v2 DSL.
Represents the failure policy of a pipeline. Currently, the default of a pipeline is that the pipeline will continue to run until no more tasks can be executed, also known as PIPELINE_FAILURE_POLICY_FAIL_SLOW. However, if a pipeline is set to PIPELINE_FAILURE_POLICY_FAIL_FAST, it will stop scheduling any new tasks when a task has failed. Any scheduled tasks will continue to completion.
The runtime artifacts of the PipelineJob. The key will be the input artifact name and the value would be one of the InputArtifact.
The type of an input artifact.
Used in:
Artifact resource id from MLMD. Which is the last portion of an artifact resource name: `projects/{project}/locations/{location}/metadataStores/default/artifacts/{artifact_id}`. The artifact must stay within the same project, location and default metadatastore as the pipeline.
The runtime detail of PipelineJob.
Used in:
Output only. The context of the pipeline.
Output only. The context of the current pipeline run.
Output only. The runtime details of the tasks under the pipeline.
Describes the state of a pipeline.
Used in:
,The pipeline state is unspecified.
The pipeline has been created or resumed, and processing has not yet begun.
The service is preparing to run the pipeline.
The pipeline is in progress.
The pipeline completed successfully.
The pipeline failed.
The pipeline is being cancelled. From this state, the pipeline may only go to either PIPELINE_STATE_SUCCEEDED, PIPELINE_STATE_FAILED or PIPELINE_STATE_CANCELLED.
The pipeline has been cancelled.
The pipeline has been stopped, and can be resumed.
The runtime detail of a task execution.
Used in:
Output only. The system generated ID of the task.
Output only. The id of the parent task if the task is within a component scope. Empty if the task is at the root level.
Output only. The user specified name of the task that is defined in [pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec].
Output only. Task create time.
Output only. Task start time.
Output only. Task end time.
Output only. The detailed execution info.
Output only. State of the task.
Output only. The execution metadata of the task.
Output only. The error that occurred during task execution. Only populated when the task's state is FAILED or CANCELLED.
Output only. A list of task status. This field keeps a record of task status evolving over time.
Output only. The runtime input artifacts of the task.
Output only. The runtime output artifacts of the task.
A list of artifact metadata.
Used in:
Output only. A list of artifact metadata.
A single record of the task status.
Used in:
Output only. Update time of this status.
Output only. The state of the task.
Output only. The error that occurred during the state. May be set when the state is any of the non-final state (PENDING/RUNNING/CANCELLING) or FAILED state. If the state is FAILED, the error here is final and not going to be retried. If the state is a non-final state, the error indicates a system-error being retried.
Specifies state of TaskExecution
Used in:
,Unspecified.
Specifies pending state for the task.
Specifies task is being executed.
Specifies task completed successfully.
Specifies Task cancel is in pending state.
Specifies task is being cancelled.
Specifies task was cancelled.
Specifies task failed.
Specifies task was skipped due to cache hit.
Specifies that the task was not triggered because the task's trigger policy is not satisfied. The trigger policy is specified in the `condition` field of [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec].
The runtime detail of a pipeline executor.
Used in:
Output only. The detailed info for a container executor.
Output only. The detailed info for a custom job executor.
The detail of a container execution. It contains the job names of the lifecycle of a container execution.
Used in:
Output only. The name of the [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the main container execution.
Output only. The name of the [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the pre-caching-check container execution. This job will be available if the [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] specifies the `pre_caching_check` hook in the lifecycle events.
Output only. The names of the previously failed [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the main container executions. The list includes the all attempts in chronological order.
Output only. The names of the previously failed [CustomJob][google.cloud.aiplatform.v1.CustomJob] for the pre-caching-check container executions. This job will be available if the [PipelineJob.pipeline_spec][google.cloud.aiplatform.v1.PipelineJob.pipeline_spec] specifies the `pre_caching_check` hook in the lifecycle events. The list includes the all attempts in chronological order.
The detailed info for a custom job executor.
Used in:
Output only. The name of the [CustomJob][google.cloud.aiplatform.v1.CustomJob].
Output only. The names of the previously failed [CustomJob][google.cloud.aiplatform.v1.CustomJob]. The list includes the all attempts in chronological order.
Pipeline template metadata if [PipelineJob.template_uri][google.cloud.aiplatform.v1.PipelineJob.template_uri] is from supported template registry. Currently, the only supported registry is Artifact Registry.
Used in:
The version_name in artifact registry. Will always be presented in output if the [PipelineJob.template_uri][google.cloud.aiplatform.v1.PipelineJob.template_uri] is from supported template registry. Format is "sha256:abcdef123456...".
Input for pointwise metric.
Used in:
Required. Spec for pointwise metric.
Required. Pointwise metric instance.
Pointwise metric instance. Usually one instance corresponds to one row in an evaluation dataset.
Used in:
Instance for pointwise metric.
Instance specified as a json string. String key-value pairs are expected in the json_instance to render PointwiseMetricSpec.instance_prompt_template.
Spec for pointwise metric result.
Used in:
Output only. Pointwise metric score.
Output only. Explanation for pointwise metric score.
Spec for pointwise metric.
Used in:
Required. Metric prompt template for pointwise metric.
Represents a network port in a container.
Used in:
The number of the port to expose on the pod's IP address. Must be a valid port number, between 1 and 65535 inclusive.
Post startup script config.
Used in:
Optional. Post startup script to run after runtime is started.
Optional. Post startup script url to download. Example: https://bucket/script.sh
Optional. Post startup script behavior that defines download and execution behavior.
Represents a notebook runtime post startup script behavior.
Used in:
Unspecified post startup script behavior.
Run post startup script after runtime is started.
Run post startup script after runtime is stopped.
Download and run post startup script every time runtime is started.
Assigns input data to training, validation, and test sets based on the value of a provided key. Supported only for tabular Datasets.
Used in:
Required. The key is a name of one of the Dataset's data columns. The value of the key (either the label's value or value in the column) must be one of {`training`, `validation`, `test`}, and it defines to which set the given piece of data is assigned. If for a piece of data the key is not present or has an invalid value, that piece is ignored by the pipeline.
Configuration for logging request-response to a BigQuery table.
Used in:
If logging is enabled or not.
Percentage of requests to be logged, expressed as a fraction in range(0,1].
BigQuery table for logging. If only given a project, a new dataset will be created with name `logging_<endpoint-display-name>_<endpoint-id>` where <endpoint-display-name> will be made BigQuery-dataset-name compatible (e.g. most special characters will become underscores). If no table name is given, a new table will be created with name `request_response_logging`
Contains the schemata used in Model's predictions and explanations via [PredictionService.Predict][google.cloud.aiplatform.v1.PredictionService.Predict], [PredictionService.Explain][google.cloud.aiplatform.v1.PredictionService.Explain] and [BatchPredictionJob][google.cloud.aiplatform.v1.BatchPredictionJob].
Used in:
, ,Immutable. Points to a YAML file stored on Google Cloud Storage describing the format of a single instance, which are used in [PredictRequest.instances][google.cloud.aiplatform.v1.PredictRequest.instances], [ExplainRequest.instances][google.cloud.aiplatform.v1.ExplainRequest.instances] and [BatchPredictionJob.input_config][google.cloud.aiplatform.v1.BatchPredictionJob.input_config]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). AutoML Models always have this field populated by Vertex AI. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Immutable. Points to a YAML file stored on Google Cloud Storage describing the parameters of prediction and explanation via [PredictRequest.parameters][google.cloud.aiplatform.v1.PredictRequest.parameters], [ExplainRequest.parameters][google.cloud.aiplatform.v1.ExplainRequest.parameters] and [BatchPredictionJob.model_parameters][google.cloud.aiplatform.v1.BatchPredictionJob.model_parameters]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). AutoML Models always have this field populated by Vertex AI, if no parameters are supported, then it is set to an empty string. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Immutable. Points to a YAML file stored on Google Cloud Storage describing the format of a single prediction produced by this Model, which are returned via [PredictResponse.predictions][google.cloud.aiplatform.v1.PredictResponse.predictions], [ExplainResponse.explanations][google.cloud.aiplatform.v1.ExplainResponse.explanations], and [BatchPredictionJob.output_config][google.cloud.aiplatform.v1.BatchPredictionJob.output_config]. The schema is defined as an OpenAPI 3.0.2 [Schema Object](https://github.com/OAI/OpenAPI-Specification/blob/main/versions/3.0.2.md#schemaObject). AutoML Models always have this field populated by Vertex AI. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Preset configuration for example-based explanations
Used in:
Preset option controlling parameters for speed-precision trade-off when querying for examples. If omitted, defaults to `PRECISE`.
The modality of the uploaded model, which automatically configures the distance measurement and feature normalization for the underlying example index and queries. If your model does not precisely fit one of these types, it is okay to choose the closest type.
Preset option controlling parameters for different modalities
Used in:
Should not be set. Added as a recommended best practice for enums
IMAGE modality
TEXT modality
TABULAR modality
Preset option controlling parameters for query speed-precision trade-off
Used in:
More precise neighbors as a trade-off against slower response.
Faster response as a trade-off against less precise neighbors.
PrivateEndpoints proto is used to provide paths for users to send requests privately. To send request via private service access, use predict_http_uri, explain_http_uri or health_http_uri. To send request via private service connect, use service_attachment.
Used in:
Output only. Http(s) path to send prediction requests.
Output only. Http(s) path to send explain requests.
Output only. Http(s) path to send health check requests.
Output only. The name of the service attachment resource. Populated if private service connect is enabled.
Represents configuration for private service connect.
Used in:
, ,Required. If true, expose the IndexEndpoint via private service connect.
A list of Projects from which the forwarding rule will target the service attachment.
Output only. The name of the generated service attachment resource. This is only populated if the endpoint is deployed with PrivateServiceConnect.
Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic.
Used in:
ExecAction probes the health of a container by executing a command.
HttpGetAction probes the health of a container by sending an HTTP GET request.
GrpcAction probes the health of a container by sending a gRPC request.
TcpSocketAction probes the health of a container by opening a TCP socket connection.
How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1. Must be less than timeout_seconds. Maps to Kubernetes probe argument 'periodSeconds'.
Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. Must be greater or equal to period_seconds. Maps to Kubernetes probe argument 'timeoutSeconds'.
Number of consecutive failures before the probe is considered failed. Defaults to 3. Minimum value is 1. Maps to Kubernetes probe argument 'failureThreshold'.
Number of consecutive successes before the probe is considered successful. Defaults to 1. Minimum value is 1. Maps to Kubernetes probe argument 'successThreshold'.
Number of seconds to wait before starting the probe. Defaults to 0. Minimum value is 0. Maps to Kubernetes probe argument 'initialDelaySeconds'.
ExecAction specifies a command to execute.
Used in:
Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.
GrpcAction checks the health of a container using a gRPC service.
Used in:
Port number of the gRPC service. Number must be in the range 1 to 65535.
Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). If this is not specified, the default behavior is defined by gRPC.
HttpGetAction describes an action based on HTTP Get requests.
Used in:
Path to access on the HTTP server.
Number of the port to access on the container. Number must be in the range 1 to 65535.
Host name to connect to, defaults to the model serving container's IP. You probably want to set "Host" in httpHeaders instead.
Scheme to use for connecting to the host. Defaults to HTTP. Acceptable values are "HTTP" or "HTTPS".
Custom headers to set in the request. HTTP allows repeated headers.
HttpHeader describes a custom header to be used in HTTP probes
Used in:
The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header.
The header field value
TcpSocketAction probes the health of a container by opening a TCP socket connection.
Used in:
Number of the port to access on the container. Number must be in the range 1 to 65535.
Optional: Host name to connect to, defaults to the model serving container's IP.
PscAutomatedEndpoints defines the output of the forwarding rule automatically created by each PscAutomationConfig.
Used in:
Corresponding project_id in pscAutomationConfigs
Corresponding network in pscAutomationConfigs.
Ip Address created by the automated forwarding rule.
Actions could take on this Publisher Model.
Used in:
Optional. To view Rest API docs.
Optional. Open notebook of the PublisherModel.
Optional. Open notebooks of the PublisherModel.
Optional. Create application using the PublisherModel.
Optional. Open fine-tuning pipeline of the PublisherModel.
Optional. Open fine-tuning pipelines of the PublisherModel.
Optional. Open prompt-tuning pipeline of the PublisherModel.
Optional. Open Genie / Playground.
Optional. Deploy the PublisherModel to Vertex Endpoint.
Optional. Deploy PublisherModel to Google Kubernetes Engine.
Optional. Open in Generation AI Studio.
Optional. Request for access.
Optional. Open evaluation pipeline of the PublisherModel.
Model metadata that is needed for UploadModel or DeployModel/CreateEndpoint requests.
Used in:
The prediction (for example, the machine) resources that the DeployedModel uses.
A description of resources that are dedicated to the DeployedModel, and that need a higher degree of manual configuration.
A description of resources that to large degree are decided by Vertex AI, and require only a modest additional configuration.
The resource name of the shared DeploymentResourcePool to deploy on. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}`
Optional. Default model display name.
Optional. Large model reference. When this is set, model_artifact_spec is not needed.
Optional. The specification of the container that is to be used when deploying this Model in Vertex AI. Not present for Large Models.
Optional. The path to the directory containing the Model artifact and any of its supporting files.
Optional. The name of the deploy task (e.g., "text to image generation").
Optional. Metadata information about this deployment config.
Required. The title of the regional resource reference.
Optional. The signed URI for ephemeral Cloud Storage access to model artifact.
Metadata information about the deployment for managing deployment config.
Used in:
Optional. Labels for the deployment. For managing deployment config like verifying, source of deployment config, etc.
Optional. Sample request for deployed endpoint.
Configurations for PublisherModel GKE deployment
Used in:
Optional. GKE deployment configuration in yaml format.
Open fine tuning pipelines.
Used in:
Required. Regional resource references to fine tuning pipelines.
Open notebooks.
Used in:
Required. Regional resource references to notebooks.
The regional resource name or the URI. Key is region, e.g., us-central1, europe-west2, global, etc..
Used in:
, ,Required.
Required.
Optional. Title of the resource.
Optional. Use case (CUJ) of the resource.
Optional. Description of the resource.
Rest API docs.
Used in:
Required.
Required. The title of the view rest API.
A named piece of documentation.
Used in:
Required. E.g., OVERVIEW, USE CASES, DOCUMENTATION, SDK & SAMPLES, JAVA, NODE.JS, etc..
Required. Content of this piece of document (in Markdown format).
An enum representing the launch stage of a PublisherModel.
Used in:
The model launch stage is unspecified.
Used to indicate the PublisherModel is at Experimental launch stage, available to a small set of customers.
Used to indicate the PublisherModel is at Private Preview launch stage, only available to a small set of customers, although a larger set of customers than an Experimental launch. Previews are the first launch stage used to get feedback from customers.
Used to indicate the PublisherModel is at Public Preview launch stage, available to all customers, although not supported for production workloads.
Used to indicate the PublisherModel is at GA launch stage, available to all customers and ready for production workload.
An enum representing the open source category of a PublisherModel.
Used in:
The open source category is unspecified, which should not be used.
Used to indicate the PublisherModel is not open sourced.
Used to indicate the PublisherModel is a Google-owned open source model w/ Google checkpoint.
Used to indicate the PublisherModel is a 3p-owned open source model w/ Google checkpoint.
Used to indicate the PublisherModel is a Google-owned pure open source model.
Used to indicate the PublisherModel is a 3p-owned pure open source model.
Reference to a resource.
Used in:
The URI of the resource.
The resource name of the Google Cloud resource.
Use case (CUJ) of the resource.
Description of the resource.
An enum representing the state of the PublicModelVersion.
Used in:
The version state is unspecified.
Used to indicate the version is stable.
Used to indicate the version is unstable.
View enumeration of PublisherModel.
Used in:
The default / unset value. The API will default to the BASIC view.
Include basic metadata about the publisher model, but not the full contents.
Include everything.
Include: VersionId, ModelVersionExternalName, and SupportedActions.
Details of operations that perform [MetadataService.PurgeArtifacts][google.cloud.aiplatform.v1.MetadataService.PurgeArtifacts].
Operation metadata for purging Artifacts.
Response message for [MetadataService.PurgeArtifacts][google.cloud.aiplatform.v1.MetadataService.PurgeArtifacts].
The number of Artifacts that this request deleted (or, if `force` is false, the number of Artifacts that will be deleted). This can be an estimate.
A sample of the Artifact names that will be deleted. Only populated if `force` is set to false. The maximum number of samples is 100 (it is possible to return fewer).
Details of operations that perform [MetadataService.PurgeContexts][google.cloud.aiplatform.v1.MetadataService.PurgeContexts].
Operation metadata for purging Contexts.
Response message for [MetadataService.PurgeContexts][google.cloud.aiplatform.v1.MetadataService.PurgeContexts].
The number of Contexts that this request deleted (or, if `force` is false, the number of Contexts that will be deleted). This can be an estimate.
A sample of the Context names that will be deleted. Only populated if `force` is set to false. The maximum number of samples is 100 (it is possible to return fewer).
Details of operations that perform [MetadataService.PurgeExecutions][google.cloud.aiplatform.v1.MetadataService.PurgeExecutions].
Operation metadata for purging Executions.
Response message for [MetadataService.PurgeExecutions][google.cloud.aiplatform.v1.MetadataService.PurgeExecutions].
The number of Executions that this request deleted (or, if `force` is false, the number of Executions that will be deleted). This can be an estimate.
A sample of the Execution names that will be deleted. Only populated if `force` is set to false. The maximum number of samples is 100 (it is possible to return fewer).
The spec of a Python packaged code.
Used in:
Required. The URI of a container image in Artifact Registry that will run the provided Python package. Vertex AI provides a wide range of executor images with pre-installed packages to meet users' various use cases. See the list of [pre-built containers for training](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). You must use an image from this list.
Required. The Google Cloud Storage location of the Python package files which are the training program and its dependent packages. The maximum number of package URIs is 100.
Required. The Python module name to run after installing the packages.
Command line arguments to be passed to the Python task.
Environment variables to be passed to the python module. Maximum limit is 100.
Input for question answering correctness metric.
Used in:
Required. Spec for question answering correctness score metric.
Required. Question answering correctness instance.
Spec for question answering correctness instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Optional. Text provided as context to answer the question.
Required. The question asked and other instruction in the inference prompt.
Spec for question answering correctness result.
Used in:
Output only. Question Answering Correctness score.
Output only. Explanation for question answering correctness score.
Output only. Confidence for question answering correctness score.
Spec for question answering correctness metric.
Used in:
Optional. Whether to use instance.reference to compute question answering correctness.
Optional. Which version to use for evaluation.
Input for question answering helpfulness metric.
Used in:
Required. Spec for question answering helpfulness score metric.
Required. Question answering helpfulness instance.
Spec for question answering helpfulness instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Optional. Text provided as context to answer the question.
Required. The question asked and other instruction in the inference prompt.
Spec for question answering helpfulness result.
Used in:
Output only. Question Answering Helpfulness score.
Output only. Explanation for question answering helpfulness score.
Output only. Confidence for question answering helpfulness score.
Spec for question answering helpfulness metric.
Used in:
Optional. Whether to use instance.reference to compute question answering helpfulness.
Optional. Which version to use for evaluation.
Input for question answering quality metric.
Used in:
Required. Spec for question answering quality score metric.
Required. Question answering quality instance.
Spec for question answering quality instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Required. Text to answer the question.
Required. Question Answering prompt for LLM.
Spec for question answering quality result.
Used in:
Output only. Question Answering Quality score.
Output only. Explanation for question answering quality score.
Output only. Confidence for question answering quality score.
Spec for question answering quality score metric.
Used in:
Optional. Whether to use instance.reference to compute question answering quality.
Optional. Which version to use for evaluation.
Input for question answering relevance metric.
Used in:
Required. Spec for question answering relevance score metric.
Required. Question answering relevance instance.
Spec for question answering relevance instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Optional. Text provided as context to answer the question.
Required. The question asked and other instruction in the inference prompt.
Spec for question answering relevance result.
Used in:
Output only. Question Answering Relevance score.
Output only. Explanation for question answering relevance score.
Output only. Confidence for question answering relevance score.
Spec for question answering relevance metric.
Used in:
Optional. Whether to use instance.reference to compute question answering relevance.
Optional. Which version to use for evaluation.
A RagChunk includes the content of a chunk of a RagFile, and associated metadata.
Used in:
, ,The content of the chunk.
If populated, represents where the chunk starts and ends in the document.
Represents where the chunk starts and ends in the document.
Used in:
Page where chunk starts in the document. Inclusive. 1-indexed.
Page where chunk ends in the document. Inclusive. 1-indexed.
Relevant contexts for one query.
Used in:
All its contexts.
A context of the query.
Used in:
If the file is imported from Cloud Storage or Google Drive, source_uri will be original file URI in Cloud Storage or Google Drive; if file is uploaded, source_uri will be file display name.
The file display name.
The text chunk.
According to the underlying Vector DB and the selected metric type, the score can be either the distance or the similarity between the query and the context and its range depends on the metric type. For example, if the metric type is COSINE_DISTANCE, it represents the distance between the query and the context. The larger the distance, the less relevant the context is to the query. The range is [0, 2], while 0 means the most relevant and 2 means the least relevant.
Context of the retrieved chunk.
A RagCorpus is a RagFile container and a project can have multiple RagCorpora.
Used as response type in: VertexRagDataService.GetRagCorpus
Used as field type in:
, ,The backend config of the RagCorpus. It can be data store and/or retrieval engine.
Optional. Immutable. The config for the Vector DBs.
Optional. Immutable. The config for the Vertex AI Search.
Output only. The resource name of the RagCorpus.
Required. The display name of the RagCorpus. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Optional. The description of the RagCorpus.
Output only. Timestamp when this RagCorpus was created.
Output only. Timestamp when this RagCorpus was last updated.
Output only. RagCorpus state.
Config for the embedding model to use for RAG.
Used in:
The model config to use.
The Vertex AI Prediction Endpoint that either refers to a publisher model or an endpoint that is hosting a 1P fine-tuned text embedding model. Endpoints hosting non-1P fine-tuned text embedding models are currently not supported. This is used for dense vector search.
Config representing a model hosted on Vertex Prediction Endpoint.
Used in:
Required. The endpoint resource name. Format: `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` or `projects/{project}/locations/{location}/endpoints/{endpoint}`
Output only. The resource name of the model that is deployed on the endpoint. Present only when the endpoint is not a publisher model. Pattern: `projects/{project}/locations/{location}/models/{model}`
Output only. Version ID of the model that is deployed on the endpoint. Present only when the endpoint is not a publisher model.
A RagFile contains user data for chunking, embedding and indexing.
Used as response type in: VertexRagDataService.GetRagFile
Used as field type in:
, ,The origin location of the RagFile if it is imported from Google Cloud Storage or Google Drive.
Output only. Google Cloud Storage location of the RagFile. It does not support wildcards in the Cloud Storage uri for now.
Output only. Google Drive location. Supports importing individual files as well as Google Drive folders.
Output only. The RagFile is encapsulated and uploaded in the UploadRagFile request.
The RagFile is imported from a Slack channel.
The RagFile is imported from a Jira query.
The RagFile is imported from a SharePoint source.
Output only. The resource name of the RagFile.
Required. The display name of the RagFile. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Optional. The description of the RagFile.
Output only. Timestamp when this RagFile was created.
Output only. Timestamp when this RagFile was last updated.
Output only. State of the RagFile.
Specifies the size and overlap of chunks for RagFiles.
Used in:
Specifies the chunking config for RagFiles.
Specifies the fixed length chunking config.
Specifies the fixed length chunking config.
Used in:
The size of the chunks.
The overlap between chunks.
Specifies the parsing config for RagFiles.
Used in:
The parser to use for RagFiles.
The Layout Parser to use for RagFiles.
The LLM Parser to use for RagFiles.
Document AI Layout Parser config.
Used in:
The full resource name of a Document AI processor or processor version. The processor must have type `LAYOUT_PARSER_PROCESSOR`. If specified, the `additional_config.parse_as_scanned_pdf` field must be false. Format: * `projects/{project_id}/locations/{location}/processors/{processor_id}` * `projects/{project_id}/locations/{location}/processors/{processor_id}/processorVersions/{processor_version_id}`
The maximum number of requests the job is allowed to make to the Document AI processor per minute. Consult https://cloud.google.com/document-ai/quotas and the Quota page for your project to set an appropriate value here. If unspecified, a default value of 120 QPM would be used.
Specifies the advanced parsing for RagFiles.
Used in:
The name of a LLM model used for parsing. Format: * `projects/{project_id}/locations/{location}/publishers/{publisher}/models/{model}`
The maximum number of requests the job is allowed to make to the LLM model per minute. Consult https://cloud.google.com/vertex-ai/generative-ai/docs/quotas and your document size to set an appropriate value here. If unspecified, a default value of 5000 QPM would be used.
The prompt to use for parsing. If not specified, a default prompt will be used.
Specifies the transformation config for RagFiles.
Used in:
,Specifies the chunking config for RagFiles.
A query to retrieve relevant contexts.
Used in:
The query to retrieve contexts. Currently only text query is supported.
Optional. The query in text format to get relevant contexts.
Optional. The retrieval config for the query.
Specifies the context retrieval config.
Used in:
,Optional. The number of contexts to retrieve.
Optional. Config for filters.
Optional. Config for ranking and reranking.
Config for filters.
Used in:
Filter contexts retrieved from the vector DB based on either vector distance or vector similarity.
Optional. Only returns contexts with vector distance smaller than the threshold.
Optional. Only returns contexts with vector similarity larger than the threshold.
Optional. String for metadata filtering.
Config for ranking and reranking.
Used in:
Config options for ranking. Currently only Rank Service is supported.
Optional. Config for Rank Service.
Optional. Config for LlmRanker.
Config for LlmRanker.
Used in:
Optional. The model name used for ranking. Format: `gemini-1.5-pro`
Config for Rank Service.
Used in:
Optional. The model name of the rank service. Format: `semantic-ranker-512@latest`
Config for the Vector DB to use for RAG.
Used in:
The config for the Vector DB.
The config for the RAG-managed Vector DB.
The config for the Pinecone.
The config for the Vertex Vector Search.
Authentication config for the chosen Vector DB.
Optional. Immutable. The embedding model config of the Vector DB.
The config for the Pinecone.
Used in:
Pinecone index name. This value cannot be changed after it's set.
The config for the default RAG-managed Vector DB.
Used in:
(message has no fields)
The config for the Vertex Vector Search.
Used in:
The resource name of the Index Endpoint. Format: `projects/{project}/locations/{location}/indexEndpoints/{index_endpoint}`
The resource name of the Index. Format: `projects/{project}/locations/{location}/indexes/{index}`
Configuration for the Ray OSS Logs.
Used in:
Optional. Flag to disable the export of Ray OSS logs to Cloud Logging.
Configuration for the Ray metrics.
Used in:
Optional. Flag to disable the Ray metrics collection.
Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes.
Used in:
Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from [Vertex prebuilt images](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field.
Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" }
Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set.
Optional. Ray metrics configurations.
Optional. OSS Ray logging configurations.
Response message for [FeaturestoreOnlineServingService.ReadFeatureValues][google.cloud.aiplatform.v1.FeaturestoreOnlineServingService.ReadFeatureValues].
Used as response type in: FeaturestoreOnlineServingService.ReadFeatureValues, FeaturestoreOnlineServingService.StreamingReadFeatureValues
Response header.
Entity view with Feature values. This may be the entity in the Featurestore if values for all Features were requested, or a projection of the entity in the Featurestore if values for only some Features were requested.
Entity view with Feature values.
Used in:
ID of the requested entity.
Each piece of data holds the k requested values for one requested Feature. If no values for the requested Feature exist, the corresponding cell will be empty. This has the same size and is in the same order as the features from the header [ReadFeatureValuesResponse.header][google.cloud.aiplatform.v1.ReadFeatureValuesResponse.header].
Container to hold value(s), successive in time, for one Feature from the request.
Used in:
Feature value if a single value is requested.
Feature values list if values, successive in time, are requested. If the requested number of values is greater than the number of existing Feature values, nonexistent values are omitted instead of being returned as empty.
Metadata for requested Features.
Used in:
Feature ID.
Response header with metadata for the requested [ReadFeatureValuesRequest.entity_type][google.cloud.aiplatform.v1.ReadFeatureValuesRequest.entity_type] and Features.
Used in:
The resource name of the EntityType from the [ReadFeatureValuesRequest][google.cloud.aiplatform.v1.ReadFeatureValuesRequest]. Value format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entityType}`.
List of Feature metadata corresponding to each piece of [ReadFeatureValuesResponse.EntityView.data][google.cloud.aiplatform.v1.ReadFeatureValuesResponse.EntityView.data].
Per month usage data
Used in:
Usage data for each user in the given month.
Per user usage data.
Used in:
User's username
Number of times the user has read data within the Tensorboard.
ReasoningEngine provides a customizable runtime for models to determine which actions to take and in which order.
Used as response type in: ReasoningEngineService.GetReasoningEngine
Used as field type in:
, ,Identifier. The resource name of the ReasoningEngine.
Required. The display name of the ReasoningEngine.
Optional. The description of the ReasoningEngine.
Optional. Configurations of the ReasoningEngine
Output only. Timestamp when this ReasoningEngine was created.
Output only. Timestamp when this ReasoningEngine was most recently updated.
Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
ReasoningEngine configurations
Used in:
Optional. User provided package spec of the ReasoningEngine. Ignored when users directly specify a deployment image through `deployment_spec.first_party_image_override`, but keeping the field_behavior to avoid introducing breaking changes.
Optional. The specification of a Reasoning Engine deployment.
Optional. Declarations for object class methods in OpenAPI specification format.
Optional. The OSS agent framework used to develop the agent. Currently supported values: "langchain", "langgraph", "ag2", "custom".
The specification of a Reasoning Engine deployment.
Used in:
Optional. Environment variables to be set with the Reasoning Engine deployment. The environment variables can be updated through the UpdateReasoningEngine API.
Optional. Environment variables where the value is a secret in Cloud Secret Manager. To use this feature, add 'Secret Manager Secret Accessor' role (roles/secretmanager.secretAccessor) to AI Platform Reasoning Engine Service Agent.
User provided package spec like pickled object and package requirements.
Used in:
Optional. The Cloud Storage URI of the pickled python object.
Optional. The Cloud Storage URI of the dependency files in tar.gz format.
Optional. The Cloud Storage URI of the `requirements.txt` file
Optional. The Python version. Currently support 3.8, 3.9, 3.10, 3.11. If not specified, default value is 3.10.
Runtime operation information for [GenAiTuningService.RebaseTunedModel][google.cloud.aiplatform.v1.GenAiTuningService.RebaseTunedModel].
The common part of the operation generic information.
Details of operations that perform reboot PersistentResource.
Operation metadata for PersistentResource.
Progress Message for Reboot LRO
A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity.
Used in:
Required. Specifies the reservation affinity type.
Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value.
Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation.
Identifies a type of reservation affinity.
Used in:
Default value. This should not be used.
Do not consume from any reserved capacity, only use on-demand.
Consume any reservation available, falling back to on-demand.
Consume from a specific reservation. When chosen, the reservation must be identified via the `key` and `values` fields.
Represents the spec of a group of resources of the same type, for example machine type, disk, and accelerators, in a PersistentResource.
Used in:
Immutable. The unique ID in a PersistentResource for referring to this resource pool. User can specify it if necessary. Otherwise, it's generated automatically.
Required. Immutable. The specification of a single machine.
Optional. The total number of machines to use for this resource pool.
Optional. Disk spec for the machine in this node pool.
Output only. The number of machines currently in use by training jobs for this resource pool. Will replace idle_replica_count.
Optional. Optional spec to configure GKE or Ray-on-Vertex autoscaling
The min/max number of replicas allowed if enabling autoscaling
Used in:
Optional. min replicas in the node pool, must be ≤ replica_count and < max_replica_count or will throw error. For autoscaling enabled Ray-on-Vertex, we allow min_replica_count of a resource_pool to be 0 to match the OSS Ray behavior(https://docs.ray.io/en/latest/cluster/vms/user-guides/configuring-autoscaling.html#cluster-config-parameters). As for Persistent Resource, the min_replica_count must be > 0, we added a corresponding validation inside CreatePersistentResourceRequestValidator.java.
Optional. max replicas in the node pool, must be ≥ replica_count and > min_replica_count or will throw error
Persistent Cluster runtime information as output
Used in:
Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" }
Configuration for the runtime on a PersistentResource instance, including but not limited to: * Service accounts used to run the workloads. * Whether to make it a dedicated Ray Cluster.
Used in:
Optional. Configure the use of workload identity on the PersistentResource
Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource.
Statistics information about resource consumption.
Used in:
Output only. The number of replica hours used. Note that many replicas may run in parallel, and additionally any given work may be queued for some time. Therefore this value is not strictly related to wall time.
Runtime operation information for [DatasetService.RestoreDatasetVersion][google.cloud.aiplatform.v1.DatasetService.RestoreDatasetVersion].
The common part of the operation metadata.
Defines a retrieval tool that model can call to access external knowledge.
Used in:
The source of the retrieval.
Set to use data source powered by Vertex AI Search.
Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService.
Optional. Deprecated. This option is no longer supported.
Retrieval config.
Used in:
The location of the user.
The language code of the user.
Metadata related to retrieval in the grounding flow.
Used in:
Optional. Score indicating how likely information from Google Search could help answer the prompt. The score is in the range `[0, 1]`, where 0 is the least likely and 1 is the most likely. This score is only populated when Google Search grounding and dynamic retrieval is enabled. It will be compared to the threshold to determine whether to trigger Google Search.
The data source for Vertex RagStore.
Used in:
Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support.
Optional. Only return contexts with vector distance smaller than the threshold.
The definition of the Rag resource.
Used in:
Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field.
Input for rouge metric.
Used in:
Required. Spec for rouge score metric.
Required. Repeated rouge instances.
Spec for rouge instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Rouge metric value for an instance.
Used in:
Output only. Rouge score.
Results for rouge metric.
Used in:
Output only. Rouge metric values.
Spec for rouge score metric - calculates the recall of n-grams in prediction as compared to reference - returns a score ranging between 0 and 1.
Used in:
Optional. Supported rouge types are rougen[1-9], rougeL, and rougeLsum.
Optional. Whether to use stemmer to compute rouge score.
Optional. Whether to split summaries while using rougeLsum.
Input for safety metric.
Used in:
Required. Spec for safety metric.
Required. Safety instance.
Spec for safety instance.
Used in:
Required. Output of the evaluated model.
Safety rating corresponding to the generated content.
Used in:
,Output only. Harm category.
Output only. Harm probability levels in the content.
Output only. Harm probability score.
Output only. Harm severity levels in the content.
Output only. Harm severity score.
Output only. Indicates whether the content was filtered out because of this rating.
Harm probability levels in the content.
Used in:
Harm probability unspecified.
Negligible level of harm.
Low level of harm.
Medium level of harm.
High level of harm.
Harm severity levels.
Used in:
Harm severity unspecified.
Negligible level of harm severity.
Low level of harm severity.
Medium level of harm severity.
High level of harm severity.
Spec for safety result.
Used in:
Output only. Safety score.
Output only. Explanation for safety score.
Output only. Confidence for safety score.
Safety settings.
Used in:
Required. Harm category.
Required. The harm block threshold.
Optional. Specify if the threshold is used for probability or severity score. If not specified, the threshold is used for probability score.
Probability vs severity.
Used in:
The harm block method is unspecified.
The harm block method uses both probability and severity scores.
The harm block method uses the probability score.
Probability based thresholds levels for blocking.
Used in:
Unspecified harm block threshold.
Block low threshold and above (i.e. block more).
Block medium threshold and above.
Block only high threshold (i.e. block less).
Block none.
Turn off the safety filter.
Spec for safety metric.
Used in:
Optional. Which version to use for evaluation.
Active learning data sampling config. For every active learning labeling iteration, it will select a batch of data based on the sampling strategy.
Used in:
Decides sample size for the initial batch. initial_batch_sample_percentage is used by default.
The percentage of data needed to be labeled in the first batch.
Decides sample size for the following batches. following_batch_sample_percentage is used by default.
The percentage of data needed to be labeled in each following batch (except the first batch).
Field to choose sampling strategy. Sampling strategy will decide which data should be selected for human labeling in every batch.
Sample strategy decides which subset of DataItems should be selected for human labeling in every batch.
Used in:
Default will be treated as UNCERTAINTY.
Sample the most uncertain data to label.
An attribution method that approximates Shapley values for features that contribute to the label being predicted. A sampling strategy is used to approximate the value rather than considering all subsets of features.
Used in:
Required. The number of feature permutations to consider when approximating the Shapley values. Valid range of its value is [1, 50], inclusively.
Sampling Strategy for logging, can be for both training and prediction dataset.
Used in:
,Random sample config. Will support more sampling strategies later.
Requests are randomly selected.
Used in:
Sample rate (0, 1]
A SavedQuery is a view of the dataset. It references a subset of annotations by problem type and filters.
Used in:
,Output only. Resource name of the SavedQuery.
Required. The user-defined name of the SavedQuery. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Some additional information about the SavedQuery.
Output only. Timestamp when this SavedQuery was created.
Output only. Timestamp when SavedQuery was last updated.
Output only. Filters on the Annotations in the dataset.
Required. Problem type of the SavedQuery. Allowed values: * IMAGE_CLASSIFICATION_SINGLE_LABEL * IMAGE_CLASSIFICATION_MULTI_LABEL * IMAGE_BOUNDING_POLY * IMAGE_BOUNDING_BOX * TEXT_CLASSIFICATION_SINGLE_LABEL * TEXT_CLASSIFICATION_MULTI_LABEL * TEXT_EXTRACTION * TEXT_SENTIMENT * VIDEO_CLASSIFICATION * VIDEO_OBJECT_TRACKING
Output only. Number of AnnotationSpecs in the context of the SavedQuery.
Used to perform a consistent read-modify-write update. If not set, a blind "overwrite" update happens.
Output only. If the Annotations belonging to the SavedQuery can be used for AutoML training.
One point viewable on a scalar metric plot.
Used in:
Value of the point at this step / timestamp.
An instance of a Schedule periodically schedules runs to make API calls based on user specified time specification and API request type.
Used as response type in: ScheduleService.CreateSchedule, ScheduleService.GetSchedule, ScheduleService.UpdateSchedule
Used as field type in:
, ,Required. The time specification to launch scheduled runs.
Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
Required. The API request template to launch the scheduled runs. User-specified ID is not supported in the request template.
Request for [PipelineService.CreatePipelineJob][google.cloud.aiplatform.v1.PipelineService.CreatePipelineJob]. CreatePipelineJobRequest.parent field is required (format: projects/{project}/locations/{location}).
Request for [NotebookService.CreateNotebookExecutionJob][google.cloud.aiplatform.v1.NotebookService.CreateNotebookExecutionJob].
Immutable. The resource name of the Schedule.
Required. User provided name of the Schedule. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Optional. Timestamp after which the first run can be scheduled. Default to Schedule create time if not specified.
Optional. Timestamp after which no new runs can be scheduled. If specified, The schedule will be completed when either end_time is reached or when scheduled_run_count >= max_run_count. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified.
Optional. Maximum run count of the schedule. If specified, The schedule will be completed when either started_run_count >= max_run_count or when end_time is reached. If not specified, new runs will keep getting scheduled until this Schedule is paused or deleted. Already scheduled runs will be allowed to complete. Unset if not specified.
Output only. The number of runs started by this schedule.
Output only. The state of this Schedule.
Output only. Timestamp when this Schedule was created.
Output only. Timestamp when this Schedule was updated.
Output only. Timestamp when this Schedule should schedule the next run. Having a next_run_time in the past means the runs are being started behind schedule.
Output only. Timestamp when this Schedule was last paused. Unset if never paused.
Output only. Timestamp when this Schedule was last resumed. Unset if never resumed from pause.
Required. Maximum number of runs that can be started concurrently for this Schedule. This is the limit for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable).
Optional. Whether new scheduled runs can be queued when max_concurrent_runs limit is reached. If set to true, new runs will be queued instead of skipped. Default to false.
Output only. Whether to backfill missed runs when the schedule is resumed from PAUSED state. If set to true, all missed runs will be scheduled. New runs will be scheduled after the backfill is complete. Default to false.
Output only. Response of the last scheduled run. This is the response for starting the scheduled requests and not the execution of the operations/jobs created by the requests (if applicable). Unset if no run has been scheduled yet.
Status of a scheduled run.
Used in:
The scheduled run time based on the user-specified schedule.
The response of the scheduled run.
Possible state of the schedule.
Used in:
Unspecified.
The Schedule is active. Runs are being scheduled on the user-specified timespec.
The schedule is paused. No new runs will be created until the schedule is resumed. Already started runs will be allowed to complete.
The Schedule is completed. No new runs will be scheduled. Already started runs will be allowed to complete. Schedules in completed state cannot be paused or resumed.
All parameters related to queuing and scheduling of custom jobs.
Used in:
Optional. The maximum job running time. The default is 7 days.
Optional. Restarts the entire CustomJob if a worker gets restarted. This feature can be used by distributed training jobs that are not resilient to workers leaving and joining a job.
Optional. This determines which type of scheduling strategy to use.
Optional. Indicates if the job should retry for internal errors after the job starts running. If true, overrides `Scheduling.restart_job_on_worker_restart` to false.
Optional. This is the maximum duration that a job will wait for the requested resources to be provisioned if the scheduling strategy is set to [Strategy.DWS_FLEX_START]. If set to 0, the job will wait indefinitely. The default is 24 hours.
Optional. This determines which type of scheduling strategy to use. Right now users have two options such as STANDARD which will use regular on demand resources to schedule the job, the other is SPOT which would leverage spot resources alongwith regular resources to schedule the job.
Used in:
Strategy will default to STANDARD.
Deprecated. Regular on-demand provisioning strategy.
Deprecated. Low cost by making potential use of spot resources.
Standard provisioning strategy uses regular on-demand resources.
Spot provisioning strategy uses spot resources.
Flex Start strategy uses DWS to queue for resources.
Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed.
Used in:
,Optional. The type of the data.
Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc
Optional. The title of the Schema.
Optional. The description of the data.
Optional. Indicates if the value may be null.
Optional. Default value of the data.
Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY.
Optional. Minimum number of the elements for Type.ARRAY.
Optional. Maximum number of the elements for Type.ARRAY.
Optional. Possible values of the element of primitive type with enum format. Examples: 1. We can define direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} 2. We can define apartment number as : {type:INTEGER, format:enum, enum:["101", "201", "301"]}
Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT.
Optional. The order of the properties. Not a standard field in open api spec. Only used to support the order of the properties.
Optional. Required properties of Type.OBJECT.
Optional. Minimum number of the properties for Type.OBJECT.
Optional. Maximum number of the properties for Type.OBJECT.
Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER
Optional. Maximum value of the Type.INTEGER and Type.NUMBER
Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING
Optional. Maximum length of the Type.STRING
Optional. Pattern of the Type.STRING to restrict a string to a regular expression.
Optional. Example of the object. Will only populated when the object is the root.
Optional. The value should be validated against any (one or more) of the subschemas in the list.
Optional. Allows indirect references between schema nodes. The value should be a valid reference to a child of the root `defs`. For example, the following schema defines a reference to a schema node named "Pet": type: object properties: pet: ref: #/defs/Pet defs: Pet: type: object properties: name: type: string The value of the "pet" property is a reference to the schema node named "Pet". See details in https://json-schema.org/understanding-json-schema/structuring
Optional. A map of definitions for use by `ref` Only allowed at the root of the schema.
Expression that allows ranking results based on annotation's property.
Used in:
Required. Saved query of the Annotation. Only Annotations belong to this saved query will be considered for ordering.
A comma-separated list of annotation fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Must also specify saved_query.
Google search entry point.
Used in:
Optional. Web content snippet that can be embedded in a web page or an app webview.
Optional. Base64 encoded JSON representing array of <search term, search url> tuple.
Stats requested for specific objective.
Used in:
If set, all attribution scores between [SearchModelDeploymentMonitoringStatsAnomaliesRequest.start_time][google.cloud.aiplatform.v1.SearchModelDeploymentMonitoringStatsAnomaliesRequest.start_time] and [SearchModelDeploymentMonitoringStatsAnomaliesRequest.end_time][google.cloud.aiplatform.v1.SearchModelDeploymentMonitoringStatsAnomaliesRequest.end_time] are fetched, and page token doesn't take effect in this case. Only used to retrieve attribution score for the top Features which has the highest attribution score in the latest monitoring run.
Represents an environment variable where the value is a secret in Cloud Secret Manager.
Used in:
Required. Name of the secret environment variable.
Required. Reference to a secret stored in the Cloud Secret Manager that will provide the value for this environment variable.
Reference to a secret stored in the Cloud Secret Manager that will provide the value for this environment variable.
Used in:
Required. The name of the secret in Cloud Secret Manager. Format: {secret_name}.
The Cloud Secret Manager secret version. Can be 'latest' for the latest version, an integer for a specific version, or a version alias.
Segment of the content.
Used in:
Output only. The index of a Part object within its parent Content object.
Output only. Start index in the given Part, measured in bytes. Offset from the start of the Part, inclusive, starting at zero.
Output only. End index in the given Part, measured in bytes. Offset from the start of the Part, exclusive, starting at zero.
Output only. The text corresponding to the segment from the response.
Configuration for the use of custom service account to run the workloads.
Used in:
Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents).
Optional. Required when all below conditions are met * `enable_custom_service_account` is true; * any runtime is specified via `ResourceRuntimeSpec` on creation time, for example, Ray The users must have `iam.serviceAccounts.actAs` permission on this service account and then the specified runtime containers will run as it. Do not set this field if you want to submit jobs using custom service account to this PersistentResource after creation, but only specify the `service_account` inside the job.
The SharePointSources to pass to ImportRagFiles.
Used in:
,The SharePoint sources.
An individual SharePointSource.
Used in:
The SharePoint folder source. If not provided, uses "root".
The path of the SharePoint folder to download from.
The ID of the SharePoint folder to download from.
The SharePoint drive source.
The name of the drive to download from.
The ID of the drive to download from.
The Application ID for the app registered in Microsoft Azure Portal. The application must also be configured with MS Graph permissions "Files.ReadAll", "Sites.ReadAll" and BrowserSiteLists.Read.All.
The application secret for the app registered in Azure.
Unique identifier of the Azure Active Directory Instance.
The name of the SharePoint site to download from. This can be the site name or the site id.
Output only. The SharePoint file id. Output only.
A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm).
Used in:
,Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails.
The Slack source for the ImportRagFilesRequest.
Used in:
,Required. The Slack channels.
SlackChannels contains the Slack channels and corresponding access token.
Used in:
Required. The Slack channel IDs.
Required. The SecretManager secret version resource name (e.g. projects/{project}/secrets/{secret}/versions/{version}) storing the Slack channel access token that has access to the slack channel IDs. See: https://api.slack.com/tutorials/tracks/getting-a-token.
SlackChannel contains the Slack channel ID and the time range to import.
Used in:
Required. The Slack channel ID.
Optional. The starting timestamp for messages to import.
Optional. The ending timestamp for messages to import.
Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
Used in:
,Represents the standard deviation of the gaussian kernel that will be used to add noise to the interpolated inputs prior to computing gradients.
This is a single float value and will be used to add noise to all the features. Use this field when all features are normalized to have the same distribution: scale to range [0, 1], [-1, 1] or z-scoring, where features are normalized to have 0-mean and 1-variance. Learn more about [normalization](https://developers.google.com/machine-learning/data-prep/transform/normalization). For best results the recommended value is about 10% - 20% of the standard deviation of the input feature. Refer to section 3.2 of the SmoothGrad paper: https://arxiv.org/pdf/1706.03825.pdf. Defaults to 0.1. If the distribution is different per feature, set [feature_noise_sigma][google.cloud.aiplatform.v1.SmoothGradConfig.feature_noise_sigma] instead for each feature.
This is similar to [noise_sigma][google.cloud.aiplatform.v1.SmoothGradConfig.noise_sigma], but provides additional flexibility. A separate noise sigma can be provided for each feature, which is useful if their distributions are different. No noise is added to features that are not set. If this field is unset, [noise_sigma][google.cloud.aiplatform.v1.SmoothGradConfig.noise_sigma] will be used for all features.
The number of gradient samples to use for approximation. The higher this number, the more accurate the gradient is, but the runtime complexity increases by this factor as well. Valid range of its value is [1, 50]. Defaults to 3.
SpecialistPool represents customers' own workforce to work on their data labeling jobs. It includes a group of specialist managers and workers. Managers are responsible for managing the workers in this pool as well as customers' data labeling jobs associated with this pool. Customers create specialist pool as well as start data labeling jobs on Cloud, managers and workers handle the jobs using CrowdCompute console.
Used as response type in: SpecialistPoolService.GetSpecialistPool
Used as field type in:
, ,Required. The resource name of the SpecialistPool.
Required. The user-defined name of the SpecialistPool. The name can be up to 128 characters long and can consist of any UTF-8 characters. This field should be unique on project-level.
Output only. The number of managers in this SpecialistPool.
The email addresses of the managers in the SpecialistPool.
Output only. The resource name of the pending data labeling jobs.
The email addresses of workers in the SpecialistPool.
Configuration for Speculative Decoding.
Used in:
The type of speculation method to use.
draft model speculation.
N-Gram speculation.
The number of speculative tokens to generate at each step.
Draft model speculation works by using the smaller model to generate candidate tokens for speculative decoding.
Used in:
Required. The resource name of the draft model.
N-Gram speculation works by trying to find matching tokens in the previous prompt sequence and use those as speculation for generating new tokens.
Used in:
The number of last N input tokens used as ngram to search/match against the previous prompt sequence. This is equal to the N in N-Gram. The default value is 3 if not specified.
Metadata information for [NotebookService.StartNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StartNotebookRuntime].
The operation generic information.
A human-readable message that shows the intermediate progress details of NotebookRuntime.
Response message for [NotebookService.StartNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StartNotebookRuntime].
(message has no fields)
Metadata information for [NotebookService.StopNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StopNotebookRuntime].
The operation generic information.
Response message for [NotebookService.StopNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.StopNotebookRuntime].
(message has no fields)
Assigns input data to the training, validation, and test sets so that the distribution of values found in the categorical column (as specified by the `key` field) is mirrored within each split. The fraction values determine the relative sizes of the splits. For example, if the specified column has three values, with 50% of the rows having value "A", 25% value "B", and 25% value "C", and the split fractions are specified as 80/10/10, then the training set will constitute 80% of the training data, with about 50% of the training set rows having the value "A" for the specified column, about 25% having the value "B", and about 25% having the value "C". Only the top 500 occurring values are used; any values not in the top 500 values are randomly assigned to a split. If less than three rows contain a specific value, those rows are randomly assigned. Supported only for tabular Datasets.
Used in:
The fraction of the input data that is to be used to train the Model.
The fraction of the input data that is to be used to validate the Model.
The fraction of the input data that is to be used to evaluate the Model.
Required. The key is a name of one of the Dataset's data columns. The key provided must be for a categorical column.
Request message for [PredictionService.StreamingPredict][google.cloud.aiplatform.v1.PredictionService.StreamingPredict]. The first message must contain [endpoint][google.cloud.aiplatform.v1.StreamingPredictRequest.endpoint] field and optionally [input][]. The subsequent messages must contain [input][].
Used as request type in: PredictionService.ServerStreamingPredict, PredictionService.StreamingPredict
Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`
The prediction input.
The parameters that govern the prediction.
Response message for [PredictionService.StreamingPredict][google.cloud.aiplatform.v1.PredictionService.StreamingPredict].
Used as response type in: PredictionService.ServerStreamingPredict, PredictionService.StreamingPredict
The prediction output.
The parameters that govern the prediction.
A list of string values.
Used in:
A list of string values.
One field of a Struct (or object) type feature value.
Used in:
Name of the field in the struct feature.
The value for this field.
Struct (or object) type feature value.
Used in:
A list of field values.
A message representing a Study.
Used as response type in: VizierService.CreateStudy, VizierService.GetStudy, VizierService.LookupStudy
Used as field type in:
,Output only. The name of a study. The study's globally unique identifier. Format: `projects/{project}/locations/{location}/studies/{study}`
Required. Describes the Study, default value is empty string.
Required. Configuration of the Study.
Output only. The detailed state of a Study.
Output only. Time at which the study was created.
Output only. A human readable reason why the Study is inactive. This should be empty if a study is ACTIVE or COMPLETED.
Describes the Study state.
Used in:
,The study state is unspecified.
The study is active.
The study is stopped due to an internal error.
The study is done when the service exhausts the parameter search space or max_trial_count is reached.
Represents specification of a Study.
Used in:
,The automated early stopping spec using decay curve rule.
The automated early stopping spec using median rule.
The automated early stopping spec using convex stopping rule.
Required. Metric specs for the Study.
Required. The set of parameters to tune.
The search algorithm specified for the Study.
The observation noise level of the study. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline.
Describe which measurement selection type will be used
Conditions for automated stopping of a Study. Enable automated stopping by configuring at least one condition.
The available search algorithms for the Study.
Used in:
The default algorithm used by Vertex AI for [hyperparameter tuning](https://cloud.google.com/vertex-ai/docs/training/hyperparameter-tuning-overview) and [Vertex AI Vizier](https://cloud.google.com/vertex-ai/docs/vizier).
Simple grid search within the feasible space. To use grid search, all parameters must be `INTEGER`, `CATEGORICAL`, or `DISCRETE`.
Simple random search within the feasible space.
Configuration for ConvexAutomatedStoppingSpec. When there are enough completed trials (configured by min_measurement_count), for pending trials with enough measurements and steps, the policy first computes an overestimate of the objective value at max_num_steps according to the slope of the incomplete objective value curve. No prediction can be made if the curve is completely flat. If the overestimation is worse than the best objective value of the completed trials, this pending trial will be early-stopped, but a last measurement will be added to the pending trial with max_num_steps and predicted objective value from the autoregression model.
Used in:
Steps used in predicting the final objective for early stopped trials. In general, it's set to be the same as the defined steps in training / tuning. If not defined, it will learn it from the completed trials. When use_steps is false, this field is set to the maximum elapsed seconds.
Minimum number of steps for a trial to complete. Trials which do not have a measurement with step_count > min_step_count won't be considered for early stopping. It's ok to set it to 0, and a trial can be early stopped at any stage. By default, min_step_count is set to be one-tenth of the max_step_count. When use_elapsed_duration is true, this field is set to the minimum elapsed seconds.
The minimal number of measurements in a Trial. Early-stopping checks will not trigger if less than min_measurement_count+1 completed trials or pending trials with less than min_measurement_count measurements. If not defined, the default value is 5.
The hyper-parameter name used in the tuning job that stands for learning rate. Leave it blank if learning rate is not in a parameter in tuning. The learning_rate is used to estimate the objective value of the ongoing trial.
This bool determines whether or not the rule is applied based on elapsed_secs or steps. If use_elapsed_duration==false, the early stopping decision is made according to the predicted objective values according to the target steps. If use_elapsed_duration==true, elapsed_secs is used instead of steps. Also, in this case, the parameters max_num_steps and min_num_steps are overloaded to contain max_elapsed_seconds and min_elapsed_seconds.
ConvexAutomatedStoppingSpec by default only updates the trials that needs to be early stopped using a newly trained auto-regressive model. When this flag is set to True, all stopped trials from the beginning are potentially updated in terms of their `final_measurement`. Also, note that the training logic of autoregressive models is different in this case. Enabling this option has shown better results and this may be the default option in the future.
The decay curve automated stopping rule builds a Gaussian Process Regressor to predict the final objective value of a Trial based on the already completed Trials and the intermediate measurements of the current Trial. Early stopping is requested for the current Trial if there is very low probability to exceed the optimal value found so far.
Used in:
True if [Measurement.elapsed_duration][google.cloud.aiplatform.v1.Measurement.elapsed_duration] is used as the x-axis of each Trials Decay Curve. Otherwise, [Measurement.step_count][google.cloud.aiplatform.v1.Measurement.step_count] will be used as the x-axis.
This indicates which measurement to use if/when the service automatically selects the final measurement from previously reported intermediate measurements. Choose this based on two considerations: A) Do you expect your measurements to monotonically improve? If so, choose LAST_MEASUREMENT. On the other hand, if you're in a situation where your system can "over-train" and you expect the performance to get better for a while but then start declining, choose BEST_MEASUREMENT. B) Are your measurements significantly noisy and/or irreproducible? If so, BEST_MEASUREMENT will tend to be over-optimistic, and it may be better to choose LAST_MEASUREMENT. If both or neither of (A) and (B) apply, it doesn't matter which selection type is chosen.
Used in:
Will be treated as LAST_MEASUREMENT.
Use the last measurement reported.
Use the best measurement reported.
The median automated stopping rule stops a pending Trial if the Trial's best objective_value is strictly below the median 'performance' of all completed Trials reported up to the Trial's last measurement. Currently, 'performance' refers to the running average of the objective values reported by the Trial in each measurement.
Used in:
True if median automated stopping rule applies on [Measurement.elapsed_duration][google.cloud.aiplatform.v1.Measurement.elapsed_duration]. It means that elapsed_duration field of latest measurement of current Trial is used to compute median objective value for each completed Trials.
Represents a metric to optimize.
Used in:
Required. The ID of the metric. Must not contain whitespaces and must be unique amongst all MetricSpecs.
Required. The optimization goal of the metric.
Used for safe search. In the case, the metric will be a safety metric. You must provide a separate metric for objective metric.
The available types of optimization goals.
Used in:
Goal Type will default to maximize.
Maximize the goal metric.
Minimize the goal metric.
Used in safe optimization to specify threshold levels and risk tolerance.
Used in:
Safety threshold (boundary value between safe and unsafe). NOTE that if you leave SafetyMetricConfig unset, a default value of 0 will be used.
Desired minimum fraction of safe trials (over total number of trials) that should be targeted by the algorithm at any time during the study (best effort). This should be between 0.0 and 1.0 and a value of 0.0 means that there is no minimum and an algorithm proceeds without targeting any specific fraction. A value of 1.0 means that the algorithm attempts to only Suggest safe Trials.
Describes the noise level of the repeated observations. "Noisy" means that the repeated observations with the same Trial parameters may lead to different metric evaluations.
Used in:
The default noise level chosen by Vertex AI.
Vertex AI assumes that the objective function is (nearly) perfectly reproducible, and will never repeat the same Trial parameters.
Vertex AI will estimate the amount of noise in metric evaluations, it may repeat the same Trial parameters more than once.
Represents a single parameter to optimize.
Used in:
,The value spec for a 'DOUBLE' parameter.
The value spec for an 'INTEGER' parameter.
The value spec for a 'CATEGORICAL' parameter.
The value spec for a 'DISCRETE' parameter.
Required. The ID of the parameter. Must not contain whitespaces and must be unique amongst all ParameterSpecs.
How the parameter should be scaled. Leave unset for `CATEGORICAL` parameters.
A conditional parameter node is active if the parameter's value matches the conditional node's parent_value_condition. If two items in conditional_parameter_specs have the same name, they must have disjoint parent_value_condition.
Value specification for a parameter in `CATEGORICAL` type.
Used in:
Required. The list of possible categories.
A default value for a `CATEGORICAL` parameter that is assumed to be a relatively good starting point. Unset value signals that there is no offered starting point. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline.
Represents a parameter spec with condition from its parent parameter.
Used in:
A set of parameter values from the parent ParameterSpec's feasible space.
The spec for matching values from a parent parameter of `DISCRETE` type.
The spec for matching values from a parent parameter of `INTEGER` type.
The spec for matching values from a parent parameter of `CATEGORICAL` type.
Required. The spec for a conditional parameter.
Represents the spec to match categorical values from parent parameter.
Used in:
Required. Matches values of the parent parameter of 'CATEGORICAL' type. All values must exist in `categorical_value_spec` of parent parameter.
Represents the spec to match discrete values from parent parameter.
Used in:
Required. Matches values of the parent parameter of 'DISCRETE' type. All values must exist in `discrete_value_spec` of parent parameter. The Epsilon of the value matching is 1e-10.
Represents the spec to match integer values from parent parameter.
Used in:
Required. Matches values of the parent parameter of 'INTEGER' type. All values must lie in `integer_value_spec` of parent parameter.
Value specification for a parameter in `DISCRETE` type.
Used in:
Required. A list of possible values. The list should be in increasing order and at least 1e-10 apart. For instance, this parameter might have possible settings of 1.5, 2.5, and 4.0. This list should not contain more than 1,000 values.
A default value for a `DISCRETE` parameter that is assumed to be a relatively good starting point. Unset value signals that there is no offered starting point. It automatically rounds to the nearest feasible discrete point. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline.
Value specification for a parameter in `DOUBLE` type.
Used in:
Required. Inclusive minimum value of the parameter.
Required. Inclusive maximum value of the parameter.
A default value for a `DOUBLE` parameter that is assumed to be a relatively good starting point. Unset value signals that there is no offered starting point. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline.
Value specification for a parameter in `INTEGER` type.
Used in:
Required. Inclusive minimum value of the parameter.
Required. Inclusive maximum value of the parameter.
A default value for an `INTEGER` parameter that is assumed to be a relatively good starting point. Unset value signals that there is no offered starting point. Currently only supported by the Vertex AI Vizier service. Not supported by HyperparameterTuningJob or TrainingPipeline.
The type of scaling that should be applied to this parameter.
Used in:
By default, no scaling is applied.
Scales the feasible space to (0, 1) linearly.
Scales the feasible space logarithmically to (0, 1). The entire feasible space must be strictly positive.
Scales the feasible space "reverse" logarithmically to (0, 1). The result is that values close to the top of the feasible space are spread out more than points near the bottom. The entire feasible space must be strictly positive.
The configuration (stopping conditions) for automated stopping of a Study. Conditions include trial budgets, time budgets, and convergence detection.
Used in:
If true, a Study enters STOPPING_ASAP whenever it would normally enters STOPPING state. The bottom line is: set to true if you want to interrupt on-going evaluations of Trials as soon as the study stopping condition is met. (Please see Study.State documentation for the source of truth).
Each "stopping rule" in this proto specifies an "if" condition. Before Vizier would generate a new suggestion, it first checks each specified stopping rule, from top to bottom in this list. Note that the first few rules (e.g. minimum_runtime_constraint, min_num_trials) will prevent other stopping rules from being evaluated until they are met. For example, setting `min_num_trials=5` and `always_stop_after= 1 hour` means that the Study will ONLY stop after it has 5 COMPLETED trials, even if more than an hour has passed since its creation. It follows the first applicable rule (whose "if" condition is satisfied) to make a stopping decision. If none of the specified rules are applicable, then Vizier decides that the study should not stop. If Vizier decides that the study should stop, the study enters STOPPING state (or STOPPING_ASAP if should_stop_asap = true). IMPORTANT: The automatic study state transition happens precisely as described above; that is, deleting trials or updating StudyConfig NEVER automatically moves the study state back to ACTIVE. If you want to _resume_ a Study that was stopped, 1) change the stopping conditions if necessary, 2) activate the study, and then 3) ask for suggestions. If the specified time or duration has not passed, do not stop the study.
If the specified time or duration has passed, stop the study.
If there are fewer than this many COMPLETED trials, do not stop the study.
If there are more than this many trials, stop the study.
If the objective value has not improved for this many consecutive trials, stop the study. WARNING: Effective only for single-objective studies.
If the objective value has not improved for this much time, stop the study. WARNING: Effective only for single-objective studies.
Time-based Constraint for Study
Used in:
Counts the wallclock time passed since the creation of this Study.
Compares the wallclock time to this time. Must use UTC timezone.
Details of operations that perform Trials suggestion.
Operation metadata for suggesting Trials.
The identifier of the client that is requesting the suggestion. If multiple SuggestTrialsRequests have the same `client_id`, the service will return the identical suggested Trial if the Trial is pending, and provide a new Trial if the last suggested Trial was completed.
Response message for [VizierService.SuggestTrials][google.cloud.aiplatform.v1.VizierService.SuggestTrials].
A list of Trials.
The state of the Study.
The time at which the operation was started.
The time at which operation processing completed.
Input for summarization helpfulness metric.
Used in:
Required. Spec for summarization helpfulness score metric.
Required. Summarization helpfulness instance.
Spec for summarization helpfulness instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Required. Text to be summarized.
Optional. Summarization prompt for LLM.
Spec for summarization helpfulness result.
Used in:
Output only. Summarization Helpfulness score.
Output only. Explanation for summarization helpfulness score.
Output only. Confidence for summarization helpfulness score.
Spec for summarization helpfulness score metric.
Used in:
Optional. Whether to use instance.reference to compute summarization helpfulness.
Optional. Which version to use for evaluation.
Input for summarization quality metric.
Used in:
Required. Spec for summarization quality score metric.
Required. Summarization quality instance.
Spec for summarization quality instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Required. Text to be summarized.
Required. Summarization prompt for LLM.
Spec for summarization quality result.
Used in:
Output only. Summarization Quality score.
Output only. Explanation for summarization quality score.
Output only. Confidence for summarization quality score.
Spec for summarization quality score metric.
Used in:
Optional. Whether to use instance.reference to compute summarization quality.
Optional. Which version to use for evaluation.
Input for summarization verbosity metric.
Used in:
Required. Spec for summarization verbosity score metric.
Required. Summarization verbosity instance.
Spec for summarization verbosity instance.
Used in:
Required. Output of the evaluated model.
Optional. Ground truth used to compare against the prediction.
Required. Text to be summarized.
Optional. Summarization prompt for LLM.
Spec for summarization verbosity result.
Used in:
Output only. Summarization Verbosity score.
Output only. Explanation for summarization verbosity score.
Output only. Confidence for summarization verbosity score.
Spec for summarization verbosity score metric.
Used in:
Optional. Whether to use instance.reference to compute summarization verbosity.
Optional. Which version to use for evaluation.
Hyperparameters for SFT.
Used in:
Optional. Number of complete passes the model makes over the entire training dataset during training.
Optional. Multiplier for adjusting the default learning rate.
Optional. Adapter size for tuning.
Supported adapter sizes for tuning.
Used in:
Adapter size is unspecified.
Adapter size 1.
Adapter size 4.
Adapter size 8.
Adapter size 16.
Tuning data statistics for Supervised Tuning.
Used in:
Output only. Number of examples in the tuning dataset.
Output only. Number of tuning characters in the tuning dataset.
Output only. Number of billable characters in the tuning dataset.
Output only. Number of billable tokens in the tuning dataset.
Output only. Number of tuning steps for this Tuning Job.
Output only. Dataset distributions for the user input tokens.
Output only. Dataset distributions for the user output tokens.
Output only. Dataset distributions for the messages per example.
Output only. Sample user messages in the training dataset uri.
The number of examples in the dataset that have been truncated by any amount.
A partial sample of the indices (starting from 1) of the truncated examples.
Dataset distribution for Supervised Tuning.
Used in:
Output only. Sum of a given population of values.
Output only. Sum of a given population of values that are billable.
Output only. The minimum of the population values.
Output only. The maximum of the population values.
Output only. The arithmetic mean of the values in the population.
Output only. The median of the values in the population.
Output only. The 5th percentile of the values in the population.
Output only. The 95th percentile of the values in the population.
Output only. Defines the histogram bucket.
Dataset bucket used to create a histogram for the distribution given a population of values.
Used in:
Output only. Number of values in the bucket.
Output only. Left bound of the bucket.
Output only. Right bound of the bucket.
Tuning Spec for Supervised Tuning for first party models.
Used in:
Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file.
Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file.
Optional. Hyperparameters for SFT.
The storage details for TFRecord output content.
Used in:
Required. Google Cloud Storage location.
A tensor value type.
Used in:
, , , , ,The data type of tensor.
Shape of the tensor.
Type specific representations that make it easy to create tensor protos in all languages. Only the representation corresponding to "dtype" can be set. The values hold the flattened representation of the tensor in row major order. [BOOL][google.cloud.aiplatform.v1.Tensor.DataType.BOOL]
[STRING][google.cloud.aiplatform.v1.Tensor.DataType.STRING]
[STRING][google.cloud.aiplatform.v1.Tensor.DataType.STRING]
[FLOAT][google.cloud.aiplatform.v1.Tensor.DataType.FLOAT]
[DOUBLE][google.cloud.aiplatform.v1.Tensor.DataType.DOUBLE]
[INT_8][google.cloud.aiplatform.v1.Tensor.DataType.INT8] [INT_16][google.cloud.aiplatform.v1.Tensor.DataType.INT16] [INT_32][google.cloud.aiplatform.v1.Tensor.DataType.INT32]
[INT64][google.cloud.aiplatform.v1.Tensor.DataType.INT64]
[UINT8][google.cloud.aiplatform.v1.Tensor.DataType.UINT8] [UINT16][google.cloud.aiplatform.v1.Tensor.DataType.UINT16] [UINT32][google.cloud.aiplatform.v1.Tensor.DataType.UINT32]
[UINT64][google.cloud.aiplatform.v1.Tensor.DataType.UINT64]
A list of tensor values.
A map of string to tensor.
Serialized raw tensor content.
Data type of the tensor.
Used in:
Not a legal value for DataType. Used to indicate a DataType field has not been set.
Data types that all computation devices are expected to be capable to support.
Tensorboard is a physical database that stores users' training metrics. A default Tensorboard is provided in each region of a Google Cloud project. If needed users can also create extra Tensorboards in their projects.
Used as response type in: TensorboardService.GetTensorboard
Used as field type in:
, ,Output only. Name of the Tensorboard. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}`
Required. User provided name of this Tensorboard.
Description of this Tensorboard.
Customer-managed encryption key spec for a Tensorboard. If set, this Tensorboard and all sub-resources of this Tensorboard will be secured by this key.
Output only. Consumer project Cloud Storage path prefix used to store blob data, which can either be a bucket or directory. Does not end with a '/'.
Output only. The number of Runs stored in this Tensorboard.
Output only. Timestamp when this Tensorboard was created.
Output only. Timestamp when this Tensorboard was last updated.
The labels with user-defined metadata to organize your Tensorboards. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Tensorboard (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Used to indicate if the TensorBoard instance is the default one. Each project & region can have at most one default TensorBoard instance. Creation of a default TensorBoard instance and updating an existing TensorBoard instance to be default will mark all other TensorBoard instances (if any) as non default.
Output only. Reserved for future use.
Output only. Reserved for future use.
One blob (e.g, image, graph) viewable on a blob metric plot.
Used in:
,Output only. A URI safe key uniquely identifying a blob. Can be used to locate the blob stored in the Cloud Storage bucket of the consumer project.
Optional. The bytes of the blob is not present unless it's returned by the ReadTensorboardBlobData endpoint.
One point viewable on a blob metric plot, but mostly just a wrapper message to work around repeated fields can't be used directly within `oneof` fields.
Used in:
List of blobs contained within the sequence.
A TensorboardExperiment is a group of TensorboardRuns, that are typically the results of a training job run, in a Tensorboard.
Used as response type in: TensorboardService.CreateTensorboardExperiment, TensorboardService.GetTensorboardExperiment, TensorboardService.UpdateTensorboardExperiment
Used as field type in:
, ,Output only. Name of the TensorboardExperiment. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}`
User provided name of this TensorboardExperiment.
Description of this TensorboardExperiment.
Output only. Timestamp when this TensorboardExperiment was created.
Output only. Timestamp when this TensorboardExperiment was last updated.
The labels with user-defined metadata to organize your TensorboardExperiment. Label keys and values cannot be longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one Dataset (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with `aiplatform.googleapis.com/` and are immutable. The following system labels exist for each Dataset: * `aiplatform.googleapis.com/dataset_metadata_schema`: output only. Its value is the [metadata_schema's][google.cloud.aiplatform.v1.Dataset.metadata_schema_uri] title.
Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Immutable. Source of the TensorboardExperiment. Example: a custom training job.
TensorboardRun maps to a specific execution of a training job with a given set of hyperparameter values, model definition, dataset, etc
Used as response type in: TensorboardService.CreateTensorboardRun, TensorboardService.GetTensorboardRun, TensorboardService.UpdateTensorboardRun
Used as field type in:
, , ,Output only. Name of the TensorboardRun. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Required. User provided name of this TensorboardRun. This value must be unique among all TensorboardRuns belonging to the same parent TensorboardExperiment.
Description of this TensorboardRun.
Output only. Timestamp when this TensorboardRun was created.
Output only. Timestamp when this TensorboardRun was last updated.
The labels with user-defined metadata to organize your TensorboardRuns. This field will be used to filter and visualize Runs in the Tensorboard UI. For example, a Vertex AI training job can set a label aiplatform.googleapis.com/training_job_id=xxxxx to all the runs created within that job. An end user can set a label experiment_id=xxxxx for all the runs produced in a Jupyter notebook. These runs can be grouped by a label value and visualized together in the Tensorboard UI. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one TensorboardRun (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
One point viewable on a tensor metric plot.
Used in:
Required. Serialized form of https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/framework/tensor.proto
Optional. Version number of TensorProto used to serialize [value][google.cloud.aiplatform.v1.TensorboardTensor.value].
TensorboardTimeSeries maps to times series produced in training runs
Used as response type in: TensorboardService.CreateTensorboardTimeSeries, TensorboardService.GetTensorboardTimeSeries, TensorboardService.UpdateTensorboardTimeSeries
Used as field type in:
, , ,Output only. Name of the TensorboardTimeSeries.
Required. User provided name of this TensorboardTimeSeries. This value should be unique among all TensorboardTimeSeries resources belonging to the same TensorboardRun resource (parent resource).
Description of this TensorboardTimeSeries.
Required. Immutable. Type of TensorboardTimeSeries value.
Output only. Timestamp when this TensorboardTimeSeries was created.
Output only. Timestamp when this TensorboardTimeSeries was last updated.
Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
Immutable. Name of the plugin this time series pertain to. Such as Scalar, Tensor, Blob
Data of the current plugin, with the size limited to 65KB.
Output only. Scalar, Tensor, or Blob metadata for this TensorboardTimeSeries.
Describes metadata for a TensorboardTimeSeries.
Used in:
Output only. Max step index of all data points within a TensorboardTimeSeries.
Output only. Max wall clock timestamp of all data points within a TensorboardTimeSeries.
Output only. The largest blob sequence length (number of blobs) of all data points in this time series, if its ValueType is BLOB_SEQUENCE.
An enum representing the value type of a TensorboardTimeSeries.
Used in:
,The value type is unspecified.
Used for TensorboardTimeSeries that is a list of scalars. E.g. accuracy of a model over epochs/time.
Used for TensorboardTimeSeries that is a list of tensors. E.g. histograms of weights of layer in a model over epoch/time.
Used for TensorboardTimeSeries that is a list of blob sequences. E.g. set of sample images with labels over epochs/time.
The config for feature monitoring threshold.
Used in:
, ,Specify a threshold value that can trigger the alert. If this threshold config is for feature distribution distance: 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. Each feature must have a non-zero threshold if they need to be monitored. Otherwise no alert will be triggered for that feature.
All the data stored in a TensorboardTimeSeries.
Used in:
, ,Required. The ID of the TensorboardTimeSeries, which will become the final component of the TensorboardTimeSeries' resource name
Required. Immutable. The value type of this time series. All the values in this time series data must match this value type.
Required. Data points in this time series.
A TensorboardTimeSeries data point.
Used in:
,Value of this time series data point.
A scalar value.
A tensor value.
A blob sequence value.
Wall clock timestamp when this data point is generated by the end user.
Step index of this data point within the run.
Assigns input data to training, validation, and test sets based on a provided timestamps. The youngest data pieces are assigned to training set, next to validation set, and the oldest to the test set. Supported only for tabular Datasets.
Used in:
The fraction of the input data that is to be used to train the Model.
The fraction of the input data that is to be used to validate the Model.
The fraction of the input data that is to be used to evaluate the Model.
Required. The key is a name of one of the Dataset's data columns. The values of the key (the values in the column) must be in RFC 3339 `date-time` format, where `time-offset` = `"Z"` (e.g. 1985-04-12T23:20:50.52Z). If for a piece of data the key is not present or has an invalid value, that piece is ignored by the pipeline.
Tokens info with a list of tokens and the corresponding list of token ids.
Used in:
A list of tokens from the input.
A list of token ids from the input.
Optional. Optional fields for the role from the corresponding Content.
Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval).
Used in:
, ,Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating [FunctionCall][google.cloud.aiplatform.v1.Part.function_call] in the response. User should provide a [FunctionResponse][google.cloud.aiplatform.v1.Part.function_response] for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided.
Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation.
Optional. GoogleSearch tool type. Tool to support Google Search in Model. Powered by Google.
Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search.
Optional. Tool to support searching public web data, powered by Vertex AI Search and Sec4 compliance.
Optional. CodeExecution tool type. Enables the model to execute code as part of generation.
Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool.
Used in:
(message has no fields)
GoogleSearch tool type. Tool to support Google Search in Model. Powered by Google.
Used in:
(message has no fields)
Input for tool call valid metric.
Used in:
Required. Spec for tool call valid metric.
Required. Repeated tool call valid instances.
Spec for tool call valid instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Tool call valid metric value for an instance.
Used in:
Output only. Tool call valid score.
Results for tool call valid metric.
Used in:
Output only. Tool call valid metric values.
Spec for tool call valid metric.
Used in:
(message has no fields)
Tool config. This config is shared for all tools provided in the request.
Used in:
,Optional. Function calling config.
Optional. Retrieval config.
Input for tool name match metric.
Used in:
Required. Spec for tool name match metric.
Required. Repeated tool name match instances.
Spec for tool name match instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Tool name match metric value for an instance.
Used in:
Output only. Tool name match score.
Results for tool name match metric.
Used in:
Output only. Tool name match metric values.
Spec for tool name match metric.
Used in:
(message has no fields)
Input for tool parameter key value match metric.
Used in:
Required. Spec for tool parameter key value match metric.
Required. Repeated tool parameter key value match instances.
Spec for tool parameter key value match instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Tool parameter key value match metric value for an instance.
Used in:
Output only. Tool parameter key value match score.
Results for tool parameter key value match metric.
Used in:
Output only. Tool parameter key value match metric values.
Spec for tool parameter key value match metric.
Used in:
Optional. Whether to use STRICT string match on parameter values.
Input for tool parameter key match metric.
Used in:
Required. Spec for tool parameter key match metric.
Required. Repeated tool parameter key match instances.
Spec for tool parameter key match instance.
Used in:
Required. Output of the evaluated model.
Required. Ground truth used to compare against the prediction.
Tool parameter key match metric value for an instance.
Used in:
Output only. Tool parameter key match score.
Results for tool parameter key match metric.
Used in:
Output only. Tool parameter key match metric values.
Spec for tool parameter key match metric.
Used in:
(message has no fields)
CMLE training config. For every active learning labeling iteration, system will train a machine learning model on CMLE. The trained model will be used by data sampling algorithm to select DataItems.
Used in:
The timeout hours for the CMLE training job, expressed in milli hours i.e. 1,000 value in this field means 1 hour.
The TrainingPipeline orchestrates tasks associated with training a Model. It always executes the training task, and optionally may also export data from Vertex AI's Dataset which becomes the training input, [upload][google.cloud.aiplatform.v1.ModelService.UploadModel] the Model to Vertex AI, and evaluate the Model.
Used as response type in: PipelineService.CreateTrainingPipeline, PipelineService.GetTrainingPipeline
Used as field type in:
,Output only. Resource name of the TrainingPipeline.
Required. The user-defined name of this TrainingPipeline.
Specifies Vertex AI owned input data that may be used for training the Model. The TrainingPipeline's [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition] should make clear whether this config is used and if there are any special requirements on how it should be filled. If nothing about this config is mentioned in the [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition], then it should be assumed that the TrainingPipeline does not depend on this configuration.
Required. A Google Cloud Storage path to the YAML file that defines the training task which is responsible for producing the model artifact, and may also include additional auxiliary work. The definition files that can be used here are found in gs://google-cloud-aiplatform/schema/trainingjob/definition/. Note: The URI given on output will be immutable and probably different, including the URI scheme, than the one given on input. The output URI will point to a location where the user only has a read access.
Required. The training task's parameter(s), as specified in the [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition]'s `inputs`.
Output only. The metadata information as specified in the [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition]'s `metadata`. This metadata is an auxiliary runtime and final information about the training task. While the pipeline is running this information is populated only at a best effort basis. Only present if the pipeline's [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition] contains `metadata` object.
Describes the Model that may be uploaded (via [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel]) by this TrainingPipeline. The TrainingPipeline's [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition] should make clear whether this Model description should be populated, and if there are any special requirements regarding how it should be filled. If nothing is mentioned in the [training_task_definition][google.cloud.aiplatform.v1.TrainingPipeline.training_task_definition], then it should be assumed that this field should not be filled and the training task either uploads the Model without a need of this information, or that training task does not support uploading a Model as part of the pipeline. When the Pipeline's state becomes `PIPELINE_STATE_SUCCEEDED` and the trained Model had been uploaded into Vertex AI, then the model_to_upload's resource [name][google.cloud.aiplatform.v1.Model.name] is populated. The Model is always uploaded into the Project and Location in which this pipeline is.
Optional. The ID to use for the uploaded Model, which will become the final component of the model resource name. This value may be up to 63 characters, and valid characters are `[a-z0-9_-]`. The first character cannot be a number or hyphen.
Optional. When specify this field, the `model_to_upload` will not be uploaded as a new model, instead, it will become a new version of this `parent_model`.
Output only. The detailed state of the pipeline.
Output only. Only populated when the pipeline's state is `PIPELINE_STATE_FAILED` or `PIPELINE_STATE_CANCELLED`.
Output only. Time when the TrainingPipeline was created.
Output only. Time when the TrainingPipeline for the first time entered the `PIPELINE_STATE_RUNNING` state.
Output only. Time when the TrainingPipeline entered any of the following states: `PIPELINE_STATE_SUCCEEDED`, `PIPELINE_STATE_FAILED`, `PIPELINE_STATE_CANCELLED`.
Output only. Time when the TrainingPipeline was most recently updated.
The labels with user-defined metadata to organize TrainingPipelines. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Customer-managed encryption key spec for a TrainingPipeline. If set, this TrainingPipeline will be secured by this key. Note: Model trained by this TrainingPipeline is also secured by this key if [model_to_upload][google.cloud.aiplatform.v1.TrainingPipeline.encryption_spec] is not set separately.
A message representing a Trial. A Trial contains a unique set of Parameters that has been or will be evaluated, along with the objective metrics got by running the Trial.
Used as response type in: VizierService.AddTrialMeasurement, VizierService.CompleteTrial, VizierService.CreateTrial, VizierService.GetTrial, VizierService.StopTrial
Used as field type in:
, , , ,Output only. Resource name of the Trial assigned by the service.
Output only. The identifier of the Trial assigned by the service.
Output only. The detailed state of the Trial.
Output only. The parameters of the Trial.
Output only. The final measurement containing the objective value.
Output only. A list of measurements that are strictly lexicographically ordered by their induced tuples (steps, elapsed_duration). These are used for early stopping computations.
Output only. Time when the Trial was started.
Output only. Time when the Trial's status changed to `SUCCEEDED` or `INFEASIBLE`.
Output only. The identifier of the client that originally requested this Trial. Each client is identified by a unique client_id. When a client asks for a suggestion, Vertex AI Vizier will assign it a Trial. The client should evaluate the Trial, complete it, and report back to Vertex AI Vizier. If suggestion is asked again by same client_id before the Trial is completed, the same Trial will be returned. Multiple clients with different client_ids can ask for suggestions simultaneously, each of them will get their own Trial.
Output only. A human readable string describing why the Trial is infeasible. This is set only if Trial state is `INFEASIBLE`.
Output only. The CustomJob name linked to the Trial. It's set for a HyperparameterTuningJob's Trial.
Output only. URIs for accessing [interactive shells](https://cloud.google.com/vertex-ai/docs/training/monitor-debug-interactive-shell) (one URI for each training node). Only available if this trial is part of a [HyperparameterTuningJob][google.cloud.aiplatform.v1.HyperparameterTuningJob] and the job's [trial_job_spec.enable_web_access][google.cloud.aiplatform.v1.CustomJobSpec.enable_web_access] field is `true`. The keys are names of each node used for the trial; for example, `workerpool0-0` for the primary node, `workerpool1-0` for the first node in the second worker pool, and `workerpool1-1` for the second node in the second worker pool. The values are the URIs for each node's interactive shell.
A message representing a parameter to be tuned.
Used in:
,Output only. The ID of the parameter. The parameter should be defined in [StudySpec's Parameters][google.cloud.aiplatform.v1.StudySpec.parameters].
Output only. The value of the parameter. `number_value` will be set if a parameter defined in StudySpec is in type 'INTEGER', 'DOUBLE' or 'DISCRETE'. `string_value` will be set if a parameter defined in StudySpec is in type 'CATEGORICAL'.
Describes a Trial state.
Used in:
The Trial state is unspecified.
Indicates that a specific Trial has been requested, but it has not yet been suggested by the service.
Indicates that the Trial has been suggested.
Indicates that the Trial should stop according to the service.
Indicates that the Trial is completed successfully.
Indicates that the Trial should not be attempted again. The service will set a Trial to INFEASIBLE when it's done but missing the final_measurement.
Used in:
A human-readable field which can store a description of this context. This will become part of the resulting Trial's description field.
If/when a Trial is generated or selected from this Context, its Parameters will match any parameters specified here. (I.e. if this context specifies parameter name:'a' int_value:3, then a resulting Trial will have int_value:3 for its parameter named 'a'.) Note that we first attempt to match existing REQUESTED Trials with contexts, and if there are no matches, we generate suggestions in the subspace defined by the parameters specified here. NOTE: a Context without any Parameters matches the entire feasible search space.
The Model Registry Model and Online Prediction Endpoint assiociated with this [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Used in:
Output only. The resource name of the TunedModel. Format: `projects/{project}/locations/{location}/models/{model}`.
Output only. A resource name of an Endpoint. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}`.
TunedModel Reference for legacy model migration.
Used in:
The Tuned Model Reference for the model.
Support migration from model registry.
Support migration from tuning job list page, from gemini-1.0-pro-002 to 1.5 and above.
Support migration from tuning job list page, from bison model to gemini model.
The tuning data statistic values for [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Used in:
The SFT Tuning data stats.
Represents a TuningJob that runs with Google owned models.
Used as response type in: GenAiTuningService.CreateTuningJob, GenAiTuningService.GetTuningJob
Used as field type in:
, ,The base model that is being tuned, e.g., "gemini-1.0-pro-002".
Tuning Spec for Supervised Fine Tuning.
Output only. Identifier. Resource name of a TuningJob. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}`
Optional. The display name of the [TunedModel][google.cloud.aiplatform.v1.Model]. The name can be up to 128 characters long and can consist of any UTF-8 characters.
Optional. The description of the [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Output only. The detailed state of the job.
Output only. Time when the [TuningJob][google.cloud.aiplatform.v1.TuningJob] was created.
Output only. Time when the [TuningJob][google.cloud.aiplatform.v1.TuningJob] for the first time entered the `JOB_STATE_RUNNING` state.
Output only. Time when the TuningJob entered any of the following [JobStates][google.cloud.aiplatform.v1.JobState]: `JOB_STATE_SUCCEEDED`, `JOB_STATE_FAILED`, `JOB_STATE_CANCELLED`, `JOB_STATE_EXPIRED`.
Output only. Time when the [TuningJob][google.cloud.aiplatform.v1.TuningJob] was most recently updated.
Output only. Only populated when job's state is `JOB_STATE_FAILED` or `JOB_STATE_CANCELLED`.
Optional. The labels with user-defined metadata to organize [TuningJob][google.cloud.aiplatform.v1.TuningJob] and generated resources such as [Model][google.cloud.aiplatform.v1.Model] and [Endpoint][google.cloud.aiplatform.v1.Endpoint]. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels.
Output only. The Experiment associated with this [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Output only. The tuned model resources assiociated with this [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Output only. The tuning data statistics associated with this [TuningJob][google.cloud.aiplatform.v1.TuningJob].
Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key.
The service account that the tuningJob workload runs as. If not specified, the Vertex AI Secure Fine-Tuned Service Agent in the project will be used. See https://cloud.google.com/iam/docs/service-agents#vertex-ai-secure-fine-tuning-service-agent Users starting the pipeline must have the `iam.serviceAccounts.actAs` permission on this service account.
Type contains the list of OpenAPI data types as defined by https://swagger.io/docs/specification/data-models/data-types/
Used in:
Not specified, should not be used.
OpenAPI string type
OpenAPI number type
OpenAPI integer type
OpenAPI boolean type
OpenAPI array type
OpenAPI object type
Runtime operation information for [IndexEndpointService.UndeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.UndeployIndex].
The operation generic information.
Response message for [IndexEndpointService.UndeployIndex][google.cloud.aiplatform.v1.IndexEndpointService.UndeployIndex].
(message has no fields)
Runtime operation information for [EndpointService.UndeployModel][google.cloud.aiplatform.v1.EndpointService.UndeployModel].
The operation generic information.
Response message for [EndpointService.UndeployModel][google.cloud.aiplatform.v1.EndpointService.UndeployModel].
(message has no fields)
Contains model information necessary to perform batch prediction without requiring a full model import.
Used in:
The path to the directory containing the Model artifact and any of its supporting files.
Contains the schemata used in Model's predictions and explanations
Input only. The specification of the container that is to be used when deploying this Model.
Runtime operation information for UpdateDeploymentResourcePool method.
The operation generic information.
Runtime operation information for [EndpointService.UpdateEndpointLongRunning][google.cloud.aiplatform.v1.EndpointService.UpdateEndpointLongRunning].
The operation generic information.
Runtime operation information for [ModelService.UpdateExplanationDataset][google.cloud.aiplatform.v1.ModelService.UpdateExplanationDataset].
The common part of the operation metadata.
Response message of [ModelService.UpdateExplanationDataset][google.cloud.aiplatform.v1.ModelService.UpdateExplanationDataset] operation.
(message has no fields)
Details of operations that perform update FeatureGroup.
Operation metadata for FeatureGroup.
Details of operations that perform update FeatureOnlineStore.
Operation metadata for FeatureOnlineStore.
Details of operations that perform update Feature.
Operation metadata for Feature Update.
Request message for [FeaturestoreService.UpdateFeature][google.cloud.aiplatform.v1.FeaturestoreService.UpdateFeature]. Request message for [FeatureRegistryService.UpdateFeature][google.cloud.aiplatform.v1.FeatureRegistryService.UpdateFeature].
Used as request type in: FeatureRegistryService.UpdateFeature, FeaturestoreService.UpdateFeature
Required. The Feature's `name` field is used to identify the Feature to be updated. Format: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}/features/{feature}` `projects/{project}/locations/{location}/featureGroups/{feature_group}/features/{feature}`
Field mask is used to specify the fields to be overwritten in the Features resource by the update. The fields specified in the update_mask are relative to the resource, not the full request. A field will be overwritten if it is in the mask. If the user does not provide a mask then only the non-empty fields present in the request will be overwritten. Set the update_mask to `*` to override all fields. Updatable fields: * `description` * `labels` * `disable_monitoring` (Not supported for FeatureRegistryService Feature) * `point_of_contact` (Not supported for FeaturestoreService FeatureStore)
Details of operations that perform update FeatureView.
Operation metadata for FeatureView Update.
Details of operations that perform update Featurestore.
Operation metadata for Featurestore.
Runtime operation information for [IndexService.UpdateIndex][google.cloud.aiplatform.v1.IndexService.UpdateIndex].
The operation generic information.
The operation metadata with regard to Matching Engine Index operation.
Runtime operation information for [JobService.UpdateModelDeploymentMonitoringJob][google.cloud.aiplatform.v1.JobService.UpdateModelDeploymentMonitoringJob].
The operation generic information.
Details of operations that perform update PersistentResource.
Operation metadata for PersistentResource.
Progress Message for Update LRO
Runtime operation information for [VertexRagDataService.UpdateRagCorpus][google.cloud.aiplatform.v1.VertexRagDataService.UpdateRagCorpus].
The operation generic information.
Details of [ReasoningEngineService.UpdateReasoningEngine][google.cloud.aiplatform.v1.ReasoningEngineService.UpdateReasoningEngine] operation.
The common part of the operation metadata.
Runtime operation metadata for [SpecialistPoolService.UpdateSpecialistPool][google.cloud.aiplatform.v1.SpecialistPoolService.UpdateSpecialistPool].
Output only. The name of the SpecialistPool to which the specialists are being added. Format: `projects/{project_id}/locations/{location_id}/specialistPools/{specialist_pool}`
The operation generic information.
Details of operations that perform update Tensorboard.
Operation metadata for Tensorboard.
Metadata information for [NotebookService.UpgradeNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.UpgradeNotebookRuntime].
The operation generic information.
A human-readable message that shows the intermediate progress details of NotebookRuntime.
Response message for [NotebookService.UpgradeNotebookRuntime][google.cloud.aiplatform.v1.NotebookService.UpgradeNotebookRuntime].
(message has no fields)
Details of [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel] operation.
The common part of the operation metadata.
Response message of [ModelService.UploadModel][google.cloud.aiplatform.v1.ModelService.UploadModel] operation.
The name of the uploaded Model resource. Format: `projects/{project}/locations/{location}/models/{model}`
Output only. The version ID of the model that is uploaded.
Config for uploading RagFile.
Used in:
Specifies the transformation config for RagFiles.
References an API call. It contains more information about long running operation and Jobs that are triggered by the API call.
Used in:
For API calls that return a long running operation. Resource name of the long running operation. Format: `projects/{project}/locations/{location}/operations/{operation}`
For API calls that start a LabelingJob. Resource name of the LabelingJob. Format: `projects/{project}/locations/{location}/dataLabelingJobs/{data_labeling_job}`
The method name of the API RPC call. For example, "/google.cloud.aiplatform.{apiVersion}.DatasetService.CreateDataset"
Value is the value of the field.
Used in:
An integer value.
A double value.
A string value.
Retrieve from Vertex AI Search datastore or engine for grounding. datastore and engine are mutually exclusive. See https://cloud.google.com/products/agent-builder
Used in:
Optional. Fully-qualified Vertex AI Search data store resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}`
Optional. Fully-qualified Vertex AI Search engine resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/engines/{engine}`
Config for the Vertex AI Search.
Used in:
Vertex AI Search Serving Config resource full name. For example, `projects/{project}/locations/{location}/collections/{collection}/engines/{engine}/servingConfigs/{serving_config}` or `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/servingConfigs/{serving_config}`.
Retrieve from Vertex RAG Store for grounding.
Used in:
,Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support.
Optional. Number of top k results to return from the selected corpora.
Optional. Only return results with vector distance smaller than the threshold.
Optional. The retrieval config for the Rag query.
The definition of the Rag resource.
Used in:
Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}`
Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field.
Metadata describes the input video content.
Used in:
Optional. The start offset of the video.
Optional. The end offset of the video.
Represents the spec of a worker pool in a job.
Used in:
The custom task to be executed in this worker pool.
The custom container task.
The Python packaged task.
Optional. Immutable. The specification of a single machine.
Optional. The number of worker replicas to use for this worker pool.
Optional. List of NFS mount spec.
Disk spec.
Contains Feature values to be written for a specific entity.
Used in:
Required. The ID of the entity.
Required. Feature values to be written, mapping from Feature ID to value. Up to 100,000 `feature_values` entries may be written across all payloads. The feature generation time, aligned by days, must be no older than five years (1825 days) and no later than one year (366 days) in the future.
Request message for [TensorboardService.WriteTensorboardRunData][google.cloud.aiplatform.v1.TensorboardService.WriteTensorboardRunData].
Used as request type in: TensorboardService.WriteTensorboardRunData
Used as field type in:
Required. The resource name of the TensorboardRun to write data to. Format: `projects/{project}/locations/{location}/tensorboards/{tensorboard}/experiments/{experiment}/runs/{run}`
Required. The TensorboardTimeSeries data to write. Values with in a time series are indexed by their step value. Repeated writes to the same step will overwrite the existing value for that step. The upper limit of data points per write request is 5000.
An explanation method that redistributes Integrated Gradients attributions to segmented regions, taking advantage of the model's fully differentiable structure. Refer to this paper for more details: https://arxiv.org/abs/1906.02825 Supported only by image Models.
Used in:
Required. The number of steps for approximating the path integral. A good value to start is 50 and gradually increase until the sum to diff property is met within the desired error range. Valid range of its value is [1, 100], inclusively.
Config for SmoothGrad approximation of gradients. When enabled, the gradients are approximated by averaging the gradients from noisy samples in the vicinity of the inputs. Adding noise can help improve the computed gradients. Refer to this paper for more details: https://arxiv.org/pdf/1706.03825.pdf
Config for XRAI with blur baseline. When enabled, a linear path from the maximally blurred image to the input image is created. Using a blurred baseline instead of zero (black image) is motivated by the BlurIG approach explained here: https://arxiv.org/abs/2004.03383