Get desktop application:
View/edit binary Protocol Buffers messages
Deprecated: Please use Dataplex Catalog instead. Data Catalog API service allows you to discover, understand, and manage your data.
Searches Data Catalog for multiple resources like entries and tags that match a query. This is a [Custom Method] (https://cloud.google.com/apis/design/custom_methods) that doesn't return all information on a resource, only its ID and high level fields. To get more information, you can subsequently call specific get methods. Note: Data Catalog search queries don't guarantee full recall. Results that match your query might not be returned, even in subsequent result pages. Additionally, returned (and not returned) results can vary if you repeat search queries. For more information, see [Data Catalog search syntax] (https://cloud.google.com/data-catalog/docs/how-to/search-reference).
Request message for [SearchCatalog][google.cloud.datacatalog.v1.DataCatalog.SearchCatalog].
Required. The scope of this search request. The `scope` is invalid if `include_org_ids`, `include_project_ids` are empty AND `include_gcp_public_datasets` is set to `false`. In this case, the request returns an error.
Optional. The query string with a minimum of 3 characters and specific syntax. For more information, see [Data Catalog search syntax](https://cloud.google.com/data-catalog/docs/how-to/search-reference). An empty query string returns all data assets (in the specified scope) that you have access to. A query string can be a simple `xyz` or qualified by predicates: * `name:x` * `column:y` * `description:z`
Upper bound on the number of results you can get in a single response. Can't be negative or 0, defaults to 10 in this case. The maximum number is 1000. If exceeded, throws an "invalid argument" exception.
Optional. Pagination token that, if specified, returns the next page of search results. If empty, returns the first page. This token is returned in the [SearchCatalogResponse.next_page_token][google.cloud.datacatalog.v1.SearchCatalogResponse.next_page_token] field of the response to a previous [SearchCatalogRequest][google.cloud.datacatalog.v1.DataCatalog.SearchCatalog] call.
Specifies the order of results. Currently supported case-sensitive values are: * `relevance` that can only be descending * `last_modified_timestamp [asc|desc]` with descending (`desc`) as default * `default` that can only be descending Search queries don't guarantee full recall. Results that match your query might not be returned, even in subsequent result pages. Additionally, returned (and not returned) results can vary if you repeat search queries. If you are experiencing recall issues and you don't have to fetch the results in any specific order, consider setting this parameter to `default`. If this parameter is omitted, it defaults to the descending `relevance`.
Optional. If set, use searchAll permission granted on organizations from `include_org_ids` and projects from `include_project_ids` instead of the fine grained per resource permissions when filtering the search results. The only allowed `order_by` criteria for admin_search mode is `default`. Using this flags guarantees a full recall of the search results.
Response message for [SearchCatalog][google.cloud.datacatalog.v1.DataCatalog.SearchCatalog].
Search results.
The approximate total number of entries matched by the query.
Pagination token that can be used in subsequent calls to retrieve the next page of results.
Unreachable locations. Search results don't include data from those locations. To get additional information on an error, repeat the search request and restrict it to specific locations by setting the `SearchCatalogRequest.scope.restricted_locations` parameter.
Creates an entry group. An entry group contains logically related entries together with [Cloud Identity and Access Management](/data-catalog/docs/concepts/iam) policies. These policies specify users who can create, edit, and view entries within entry groups. Data Catalog automatically creates entry groups with names that start with the `@` symbol for the following resources: * BigQuery entries (`@bigquery`) * Pub/Sub topics (`@pubsub`) * Dataproc Metastore services (`@dataproc_metastore_{SERVICE_NAME_HASH}`) You can create your own entry groups for Cloud Storage fileset entries and custom entries together with the corresponding IAM policies. User-created entry groups can't contain the `@` symbol, it is reserved for automatically created groups. Entry groups, like entries, can be searched. A maximum of 10,000 entry groups may be created per organization across all locations. You must enable the Data Catalog API in the project identified by the `parent` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [CreateEntryGroup][google.cloud.datacatalog.v1.DataCatalog.CreateEntryGroup].
Required. The names of the project and location that the new entry group belongs to. Note: The entry group itself and its child resources might not be stored in the location specified in its name.
Required. The ID of the entry group to create. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), and must start with a letter or underscore. The maximum size is 64 bytes when encoded in UTF-8.
The entry group to create. Defaults to empty.
Gets an entry group.
Request message for [GetEntryGroup][google.cloud.datacatalog.v1.DataCatalog.GetEntryGroup].
Required. The name of the entry group to get.
The fields to return. If empty or omitted, all fields are returned.
Updates an entry group. You must enable the Data Catalog API in the project identified by the `entry_group.name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [UpdateEntryGroup][google.cloud.datacatalog.v1.DataCatalog.UpdateEntryGroup].
Required. Updates for the entry group. The `name` field must be set.
Names of fields whose values to overwrite on an entry group. If this parameter is absent or empty, all modifiable fields are overwritten. If such fields are non-required and omitted in the request body, their values are emptied.
Deletes an entry group. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [DeleteEntryGroup][google.cloud.datacatalog.v1.DataCatalog.DeleteEntryGroup].
Required. The name of the entry group to delete.
Optional. If true, deletes all entries in the entry group.
Lists entry groups.
Request message for [ListEntryGroups][google.cloud.datacatalog.v1.DataCatalog.ListEntryGroups].
Required. The name of the location that contains the entry groups to list. Can be provided as a URL.
Optional. The maximum number of items to return. Default is 10. Maximum limit is 1000. Throws an invalid argument if `page_size` is greater than 1000.
Optional. Pagination token that specifies the next page to return. If empty, returns the first page.
Response message for [ListEntryGroups][google.cloud.datacatalog.v1.DataCatalog.ListEntryGroups].
Entry group details.
Pagination token to specify in the next call to retrieve the next page of results. Empty if there are no more items.
Creates an entry. You can create entries only with 'FILESET', 'CLUSTER', 'DATA_STREAM', or custom types. Data Catalog automatically creates entries with other types during metadata ingestion from integrated systems. You must enable the Data Catalog API in the project identified by the `parent` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project). An entry group can have a maximum of 100,000 entries.
Request message for [CreateEntry][google.cloud.datacatalog.v1.DataCatalog.CreateEntry].
Required. The name of the entry group this entry belongs to. Note: The entry itself and its child resources might not be stored in the location specified in its name.
Required. The ID of the entry to create. The ID must contain only letters (a-z, A-Z), numbers (0-9), and underscores (_). The maximum size is 64 bytes when encoded in UTF-8.
Required. The entry to create.
Updates an existing entry. You must enable the Data Catalog API in the project identified by the `entry.name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [UpdateEntry][google.cloud.datacatalog.v1.DataCatalog.UpdateEntry].
Required. Updates for the entry. The `name` field must be set.
Names of fields whose values to overwrite on an entry. If this parameter is absent or empty, all modifiable fields are overwritten. If such fields are non-required and omitted in the request body, their values are emptied. You can modify only the fields listed below. For entries with type `DATA_STREAM`: * `schema` For entries with type `FILESET`: * `schema` * `display_name` * `description` * `gcs_fileset_spec` * `gcs_fileset_spec.file_patterns` For entries with `user_specified_type`: * `schema` * `display_name` * `description` * `user_specified_type` * `user_specified_system` * `linked_resource` * `source_system_timestamps`
Deletes an existing entry. You can delete only the entries created by the [CreateEntry][google.cloud.datacatalog.v1.DataCatalog.CreateEntry] method. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [DeleteEntry][google.cloud.datacatalog.v1.DataCatalog.DeleteEntry].
Required. The name of the entry to delete.
Gets an entry.
Request message for [GetEntry][google.cloud.datacatalog.v1.DataCatalog.GetEntry].
Required. The name of the entry to get.
Gets an entry by its target resource name. The resource name comes from the source Google Cloud Platform service.
Request message for [LookupEntry][google.cloud.datacatalog.v1.DataCatalog.LookupEntry].
Required. A full name, SQL name, or a fully qualified name of a Google Cloud Platform resource.
The full name of the Google Cloud Platform resource the Data Catalog entry represents. For more information, see [Full Resource Name] (https://cloud.google.com/apis/design/resource_names#full_resource_name). Full names are case-sensitive. For example: * `//bigquery.googleapis.com/projects/{PROJECT_ID}/datasets/{DATASET_ID}/tables/{TABLE_ID}` * `//pubsub.googleapis.com/projects/{PROJECT_ID}/topics/{TOPIC_ID}`
The SQL name of the entry. SQL names are case-sensitive. Examples: * `pubsub.topic.{PROJECT_ID}.{TOPIC_ID}` * `pubsub.topic.{PROJECT_ID}.`\``{TOPIC.ID.SEPARATED.WITH.DOTS}`\` * `bigquery.table.{PROJECT_ID}.{DATASET_ID}.{TABLE_ID}` * `bigquery.dataset.{PROJECT_ID}.{DATASET_ID}` * `datacatalog.entry.{PROJECT_ID}.{LOCATION_ID}.{ENTRY_GROUP_ID}.{ENTRY_ID}` Identifiers (`*_ID`) should comply with the [Lexical structure in GoogleSQL] (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical).
[Fully Qualified Name (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names) of the resource. FQNs take two forms: * For non-regionalized resources: `{SYSTEM}:{PROJECT}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` * For regionalized resources: `{SYSTEM}:{PROJECT}.{LOCATION_ID}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` Example for a DPMS table: `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}`
Project where the lookup should be performed. Required to lookup entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system` using its `fully_qualified_name`. Ignored in other cases.
Location where the lookup should be performed. Required to lookup entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system` using its `fully_qualified_name`. Ignored in other cases.
Lists entries. Note: Currently, this method can list only custom entries. To get a list of both custom and automatically created entries, use [SearchCatalog][google.cloud.datacatalog.v1.DataCatalog.SearchCatalog].
Request message for [ListEntries][google.cloud.datacatalog.v1.DataCatalog.ListEntries].
Required. The name of the entry group that contains the entries to list. Can be provided in URL format.
The maximum number of items to return. Default is 10. Maximum limit is 1000. Throws an invalid argument if `page_size` is more than 1000.
Pagination token that specifies the next page to return. If empty, the first page is returned.
The fields to return for each entry. If empty or omitted, all fields are returned. For example, to return a list of entries with only the `name` field, set `read_mask` to only one path with the `name` value.
Response message for [ListEntries][google.cloud.datacatalog.v1.DataCatalog.ListEntries].
Entry details.
Pagination token of the next results page. Empty if there are no more items in results.
Modifies entry overview, part of the business context of an [Entry][google.cloud.datacatalog.v1.Entry]. To call this method, you must have the `datacatalog.entries.updateOverview` IAM permission on the corresponding project.
Request message for [ModifyEntryOverview][google.cloud.datacatalog.v1.DataCatalog.ModifyEntryOverview].
Required. The full resource name of the entry.
Required. The new value for the Entry Overview.
Modifies contacts, part of the business context of an [Entry][google.cloud.datacatalog.v1.Entry]. To call this method, you must have the `datacatalog.entries.updateContacts` IAM permission on the corresponding project.
Request message for [ModifyEntryContacts][google.cloud.datacatalog.v1.DataCatalog.ModifyEntryContacts].
Required. The full resource name of the entry.
Required. The new value for the Contacts.
Creates a tag template. You must enable the Data Catalog API in the project identified by the `parent` parameter. For more information, see [Data Catalog resource project] (https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [CreateTagTemplate][google.cloud.datacatalog.v1.DataCatalog.CreateTagTemplate].
Required. The name of the project and the template location [region](https://cloud.google.com/data-catalog/docs/concepts/regions).
Required. The ID of the tag template to create. The ID must contain only lowercase letters (a-z), numbers (0-9), or underscores (_), and must start with a letter or underscore. The maximum size is 64 bytes when encoded in UTF-8.
Required. The tag template to create.
Gets a tag template.
Request message for [GetTagTemplate][google.cloud.datacatalog.v1.DataCatalog.GetTagTemplate].
Required. The name of the tag template to get.
Updates a tag template. You can't update template fields with this method. These fields are separate resources with their own create, update, and delete methods. You must enable the Data Catalog API in the project identified by the `tag_template.name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [UpdateTagTemplate][google.cloud.datacatalog.v1.DataCatalog.UpdateTagTemplate].
Required. The template to update. The `name` field must be set.
Names of fields whose values to overwrite on a tag template. Currently, only `display_name` and `is_publicly_readable` can be overwritten. If this parameter is absent or empty, all modifiable fields are overwritten. If such fields are non-required and omitted in the request body, their values are emptied. Note: Updating the `is_publicly_readable` field may require up to 12 hours to take effect in search results.
Deletes a tag template and all tags that use it. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [DeleteTagTemplate][google.cloud.datacatalog.v1.DataCatalog.DeleteTagTemplate].
Required. The name of the tag template to delete.
Required. If true, deletes all tags that use this template. Currently, `true` is the only supported value.
Creates a field in a tag template. You must enable the Data Catalog API in the project identified by the `parent` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [CreateTagTemplateField][google.cloud.datacatalog.v1.DataCatalog.CreateTagTemplateField].
Required. The name of the project and the template location [region](https://cloud.google.com/data-catalog/docs/concepts/regions).
Required. The ID of the tag template field to create. Note: Adding a required field to an existing template is *not* allowed. Field IDs can contain letters (both uppercase and lowercase), numbers (0-9), underscores (_) and dashes (-). Field IDs must be at least 1 character long and at most 128 characters long. Field IDs must also be unique within their template.
Required. The tag template field to create.
Updates a field in a tag template. You can't update the field type with this method. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [UpdateTagTemplateField][google.cloud.datacatalog.v1.DataCatalog.UpdateTagTemplateField].
Required. The name of the tag template field.
Required. The template to update.
Optional. Names of fields whose values to overwrite on an individual field of a tag template. The following fields are modifiable: * `display_name` * `type.enum_type` * `is_required` If this parameter is absent or empty, all modifiable fields are overwritten. If such fields are non-required and omitted in the request body, their values are emptied with one exception: when updating an enum type, the provided values are merged with the existing values. Therefore, enum values can only be added, existing enum values cannot be deleted or renamed. Additionally, updating a template field from optional to required is *not* allowed.
Renames a field in a tag template. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project] (https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [RenameTagTemplateField][google.cloud.datacatalog.v1.DataCatalog.RenameTagTemplateField].
Required. The name of the tag template field.
Required. The new ID of this tag template field. For example, `my_new_field`.
Renames an enum value in a tag template. Within a single enum field, enum values must be unique.
Request message for [RenameTagTemplateFieldEnumValue][google.cloud.datacatalog.v1.DataCatalog.RenameTagTemplateFieldEnumValue].
Required. The name of the enum field value.
Required. The new display name of the enum value. For example, `my_new_enum_value`.
Deletes a field in a tag template and all uses of this field from the tags based on this template. You must enable the Data Catalog API in the project identified by the `name` parameter. For more information, see [Data Catalog resource project](https://cloud.google.com/data-catalog/docs/concepts/resource-project).
Request message for [DeleteTagTemplateField][google.cloud.datacatalog.v1.DataCatalog.DeleteTagTemplateField].
Required. The name of the tag template field to delete.
Required. If true, deletes this field from any tags that use it. Currently, `true` is the only supported value.
Creates a tag and assigns it to: * An [Entry][google.cloud.datacatalog.v1.Entry] if the method name is `projects.locations.entryGroups.entries.tags.create`. * Or [EntryGroup][google.cloud.datacatalog.v1.EntryGroup]if the method name is `projects.locations.entryGroups.tags.create`. Note: The project identified by the `parent` parameter for the [tag] (https://cloud.google.com/data-catalog/docs/reference/rest/v1/projects.locations.entryGroups.entries.tags/create#path-parameters) and the [tag template] (https://cloud.google.com/data-catalog/docs/reference/rest/v1/projects.locations.tagTemplates/create#path-parameters) used to create the tag must be in the same organization.
Request message for [CreateTag][google.cloud.datacatalog.v1.DataCatalog.CreateTag].
Required. The name of the resource to attach this tag to. Tags can be attached to entries or entry groups. An entry can have up to 1000 attached tags. Note: The tag and its child resources might not be stored in the location specified in its name.
Required. The tag to create.
Updates an existing tag.
Request message for [UpdateTag][google.cloud.datacatalog.v1.DataCatalog.UpdateTag].
Required. The updated tag. The "name" field must be set.
Names of fields whose values to overwrite on a tag. Currently, a tag has the only modifiable field with the name `fields`. In general, if this parameter is absent or empty, all modifiable fields are overwritten. If such fields are non-required and omitted in the request body, their values are emptied.
Deletes a tag.
Request message for [DeleteTag][google.cloud.datacatalog.v1.DataCatalog.DeleteTag].
Required. The name of the tag to delete.
Lists tags assigned to an [Entry][google.cloud.datacatalog.v1.Entry]. The [columns][google.cloud.datacatalog.v1.Tag.column] in the response are lowercased.
Request message for [ListTags][google.cloud.datacatalog.v1.DataCatalog.ListTags].
Required. The name of the Data Catalog resource to list the tags of. The resource can be an [Entry][google.cloud.datacatalog.v1.Entry] or an [EntryGroup][google.cloud.datacatalog.v1.EntryGroup] (without `/entries/{entries}` at the end).
The maximum number of tags to return. Default is 10. Maximum limit is 1000.
Pagination token that specifies the next page to return. If empty, the first page is returned.
Response message for [ListTags][google.cloud.datacatalog.v1.DataCatalog.ListTags].
[Tag][google.cloud.datacatalog.v1.Tag] details.
Pagination token of the next results page. Empty if there are no more items in results.
`ReconcileTags` creates or updates a list of tags on the entry. If the [ReconcileTagsRequest.force_delete_missing][google.cloud.datacatalog.v1.ReconcileTagsRequest.force_delete_missing] parameter is set, the operation deletes tags not included in the input tag list. `ReconcileTags` returns a [long-running operation] [google.longrunning.Operation] resource that can be queried with [Operations.GetOperation][google.longrunning.Operations.GetOperation] to return [ReconcileTagsMetadata] [google.cloud.datacatalog.v1.ReconcileTagsMetadata] and a [ReconcileTagsResponse] [google.cloud.datacatalog.v1.ReconcileTagsResponse] message.
Request message for [ReconcileTags][google.cloud.datacatalog.v1.DataCatalog.ReconcileTags].
Required. Name of [Entry][google.cloud.datacatalog.v1.Entry] to be tagged.
Required. The name of the tag template, which is used for reconciliation.
If set to `true`, deletes entry tags related to a tag template not listed in the tags source from an entry. If set to `false`, unlisted tags are retained.
A list of tags to apply to an entry. A tag can specify a tag template, which must be the template specified in the `ReconcileTagsRequest`. The sole entry and each of its columns must be mentioned at most once.
Marks an [Entry][google.cloud.datacatalog.v1.Entry] as starred by the current user. Starring information is private to each user.
Request message for [StarEntry][google.cloud.datacatalog.v1.DataCatalog.StarEntry].
Required. The name of the entry to mark as starred.
Response message for [StarEntry][google.cloud.datacatalog.v1.DataCatalog.StarEntry]. Empty for now
(message has no fields)
Marks an [Entry][google.cloud.datacatalog.v1.Entry] as NOT starred by the current user. Starring information is private to each user.
Request message for [UnstarEntry][google.cloud.datacatalog.v1.DataCatalog.UnstarEntry].
Required. The name of the entry to mark as **not** starred.
Response message for [UnstarEntry][google.cloud.datacatalog.v1.DataCatalog.UnstarEntry]. Empty for now
(message has no fields)
Sets an access control policy for a resource. Replaces any existing policy. Supported resources are: - Tag templates - Entry groups Note: This method sets policies only within Data Catalog and can't be used to manage policies in BigQuery, Pub/Sub, Dataproc Metastore, and any external Google Cloud Platform resources synced with the Data Catalog. To call this method, you must have the following Google IAM permissions: - `datacatalog.tagTemplates.setIamPolicy` to set policies on tag templates. - `datacatalog.entryGroups.setIamPolicy` to set policies on entry groups.
Gets the access control policy for a resource. May return: * A`NOT_FOUND` error if the resource doesn't exist or you don't have the permission to view it. * An empty policy if the resource exists but doesn't have a set policy. Supported resources are: - Tag templates - Entry groups Note: This method doesn't get policies from Google Cloud Platform resources ingested into Data Catalog. To call this method, you must have the following Google IAM permissions: - `datacatalog.tagTemplates.getIamPolicy` to get policies on tag templates. - `datacatalog.entryGroups.getIamPolicy` to get policies on entry groups.
Gets your permissions on a resource. Returns an empty set of permissions if the resource doesn't exist. Supported resources are: - Tag templates - Entry groups Note: This method gets policies only within Data Catalog and can't be used to get policies from BigQuery, Pub/Sub, Dataproc Metastore, and any external Google Cloud Platform resources ingested into Data Catalog. No Google IAM permissions are required to call this method.
Imports entries from a source, such as data previously dumped into a Cloud Storage bucket, into Data Catalog. Import of entries is a sync operation that reconciles the state of the third-party system with the Data Catalog. `ImportEntries` accepts source data snapshots of a third-party system. Snapshot should be delivered as a .wire or base65-encoded .txt file containing a sequence of Protocol Buffer messages of [DumpItem][google.cloud.datacatalog.v1.DumpItem] type. `ImportEntries` returns a [long-running operation] [google.longrunning.Operation] resource that can be queried with [Operations.GetOperation][google.longrunning.Operations.GetOperation] to return [ImportEntriesMetadata][google.cloud.datacatalog.v1.ImportEntriesMetadata] and an [ImportEntriesResponse][google.cloud.datacatalog.v1.ImportEntriesResponse] message.
Request message for [ImportEntries][google.cloud.datacatalog.v1.DataCatalog.ImportEntries] method.
Required. Target entry group for ingested entries.
Source of imported entries, e.g. dump stored in a Cloud Storage
Path to a Cloud Storage bucket that contains a dump ready for ingestion.
Optional. (Optional) Dataplex task job id, if specified will be used as part of ImportEntries LRO ID
Sets the configuration related to the migration to Dataplex for an organization or project.
Request message for [SetConfig][google.cloud.datacatalog.v1.DataCatalog.SetConfig].
Required. The organization or project whose config is being specified.
The configuration field to set.
Opt-in status for the migration of Tag Templates to Dataplex.
Opt-in status for the UI switch to Dataplex.
Retrieves the configuration related to the migration from Data Catalog to Dataplex for a specific organization, including all the projects under it which have a separate configuration set.
Request message for [RetrieveConfig][google.cloud.datacatalog.v1.DataCatalog.RetrieveConfig].
Required. The organization whose config is being retrieved.
The configuration related to the migration from Data Catalog to Dataplex that has been applied to an organization and any projects under it. It is the response message for [RetrieveConfig][google.cloud.datacatalog.v1.DataCatalog.RetrieveConfig].
Map of organizations and project resource names and their configuration. The format for the map keys is `organizations/{organizationId}` or `projects/{projectId}`.
Retrieves the effective configuration related to the migration from Data Catalog to Dataplex for a specific organization or project. If there is no specific configuration set for the resource, the setting is checked hierarchicahlly through the ancestors of the resource, starting from the resource itself.
Request message for [RetrieveEffectiveConfig][google.cloud.datacatalog.v1.DataCatalog.RetrieveEffectiveConfig].
Required. The resource whose effective config is being retrieved.
Policy Tag Manager API service allows you to manage your policy tags and taxonomies. Policy tags are used to tag BigQuery columns and apply additional access control policies. A taxonomy is a hierarchical grouping of policy tags that classify data along a common axis.
Creates a taxonomy in a specified project. The taxonomy is initially empty, that is, it doesn't contain policy tags.
Request message for [CreateTaxonomy][google.cloud.datacatalog.v1.PolicyTagManager.CreateTaxonomy].
Required. Resource name of the project that the taxonomy will belong to.
The taxonomy to create.
Deletes a taxonomy, including all policy tags in this taxonomy, their associated policies, and the policy tags references from BigQuery columns.
Request message for [DeleteTaxonomy][google.cloud.datacatalog.v1.PolicyTagManager.DeleteTaxonomy].
Required. Resource name of the taxonomy to delete. Note: All policy tags in this taxonomy are also deleted.
Updates a taxonomy, including its display name, description, and activated policy types.
Request message for [UpdateTaxonomy][google.cloud.datacatalog.v1.PolicyTagManager.UpdateTaxonomy].
The taxonomy to update. You can update only its description, display name, and activated policy types.
Specifies fields to update. If not set, defaults to all fields you can update. For more information, see [FieldMask] (https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#fieldmask).
Lists all taxonomies in a project in a particular location that you have a permission to view.
Request message for [ListTaxonomies][google.cloud.datacatalog.v1.PolicyTagManager.ListTaxonomies].
Required. Resource name of the project to list the taxonomies of.
The maximum number of items to return. Must be a value between 1 and 1000 inclusively. If not set, defaults to 50.
The pagination token of the next results page. If not set, the first page is returned. The token is returned in the response to a previous list request.
Supported field for filter is 'service' and value is 'dataplex'. Eg: service=dataplex.
Response message for [ListTaxonomies][google.cloud.datacatalog.v1.PolicyTagManager.ListTaxonomies].
Taxonomies that the project contains.
Pagination token of the next results page. Empty if there are no more results in the list.
Gets a taxonomy.
Request message for [GetTaxonomy][google.cloud.datacatalog.v1.PolicyTagManager.GetTaxonomy].
Required. Resource name of the taxonomy to get.
Creates a policy tag in a taxonomy.
Request message for [CreatePolicyTag][google.cloud.datacatalog.v1.PolicyTagManager.CreatePolicyTag].
Required. Resource name of the taxonomy that the policy tag will belong to.
The policy tag to create.
Deletes a policy tag together with the following: * All of its descendant policy tags, if any * Policies associated with the policy tag and its descendants * References from BigQuery table schema of the policy tag and its descendants
Request message for [DeletePolicyTag][google.cloud.datacatalog.v1.PolicyTagManager.DeletePolicyTag].
Required. Resource name of the policy tag to delete. Note: All of its descendant policy tags are also deleted.
Updates a policy tag, including its display name, description, and parent policy tag.
Request message for [UpdatePolicyTag][google.cloud.datacatalog.v1.PolicyTagManager.UpdatePolicyTag].
The policy tag to update. You can update only its description, display name, and parent policy tag fields.
Specifies the fields to update. You can update only display name, description, and parent policy tag. If not set, defaults to all updatable fields. For more information, see [FieldMask] (https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#fieldmask).
Lists all policy tags in a taxonomy.
Request message for [ListPolicyTags][google.cloud.datacatalog.v1.PolicyTagManager.ListPolicyTags].
Required. Resource name of the taxonomy to list the policy tags of.
The maximum number of items to return. Must be a value between 1 and 1000 inclusively. If not set, defaults to 50.
The pagination token of the next results page. If not set, returns the first page. The token is returned in the response to a previous list request.
Response message for [ListPolicyTags][google.cloud.datacatalog.v1.PolicyTagManager.ListPolicyTags].
The policy tags that belong to the taxonomy.
Pagination token of the next results page. Empty if there are no more results in the list.
Gets a policy tag.
Request message for [GetPolicyTag][google.cloud.datacatalog.v1.PolicyTagManager.GetPolicyTag].
Required. Resource name of the policy tag.
Gets the IAM policy for a policy tag or a taxonomy.
Sets the IAM policy for a policy tag or a taxonomy.
Returns your permissions on a specified policy tag or taxonomy.
Policy Tag Manager Serialization API service allows you to manipulate your policy tags and taxonomies in a serialized format. Taxonomy is a hierarchical group of policy tags.
Replaces (updates) a taxonomy and all its policy tags. The taxonomy and its entire hierarchy of policy tags must be represented literally by `SerializedTaxonomy` and the nested `SerializedPolicyTag` messages. This operation automatically does the following: - Deletes the existing policy tags that are missing from the `SerializedPolicyTag`. - Creates policy tags that don't have resource names. They are considered new. - Updates policy tags with valid resources names accordingly.
Request message for [ReplaceTaxonomy][google.cloud.datacatalog.v1.PolicyTagManagerSerialization.ReplaceTaxonomy].
Required. Resource name of the taxonomy to update.
Required. Taxonomy to update along with its child policy tags.
Creates new taxonomies (including their policy tags) in a given project by importing from inlined or cross-regional sources. For a cross-regional source, new taxonomies are created by copying from a source in another region. For an inlined source, taxonomies and policy tags are created in bulk using nested protocol buffer structures.
Request message for [ImportTaxonomies][google.cloud.datacatalog.v1.PolicyTagManagerSerialization.ImportTaxonomies].
Required. Resource name of project that the imported taxonomies will belong to.
Source taxonomies to import.
Inline source taxonomy to import.
Cross-regional source taxonomy to import.
Response message for [ImportTaxonomies][google.cloud.datacatalog.v1.PolicyTagManagerSerialization.ImportTaxonomies].
Imported taxonomies.
Exports taxonomies in the requested type and returns them, including their policy tags. The requested taxonomies must belong to the same project. This method generates `SerializedTaxonomy` protocol buffers with nested policy tags that can be used as input for `ImportTaxonomies` calls.
Request message for [ExportTaxonomies][google.cloud.datacatalog.v1.PolicyTagManagerSerialization.ExportTaxonomies].
Required. Resource name of the project that the exported taxonomies belong to.
Required. Resource names of the taxonomies to export.
Required. Export destination for taxonomies.
Serialized export taxonomies that contain all the policy tags as nested protocol buffers.
Response message for [ExportTaxonomies][google.cloud.datacatalog.v1.PolicyTagManagerSerialization.ExportTaxonomies].
List of taxonomies and policy tags as nested protocol buffers.
Specification for the BigQuery connection.
Used in:
The type of the BigQuery connection.
Specification for the BigQuery connection to a Cloud SQL instance.
True if there are credentials attached to the BigQuery connection; false otherwise.
The type of the BigQuery connection.
Used in:
Unspecified type.
Cloud SQL connection.
Specification for a group of BigQuery tables with the `[prefix]YYYYMMDD` name pattern. For more information, see [Introduction to partitioned tables] (https://cloud.google.com/bigquery/docs/partitioned-tables#partitioning_versus_sharding).
Used in:
Output only. The Data Catalog resource name of the dataset entry the current table belongs to. For example: `projects/{PROJECT_ID}/locations/{LOCATION}/entrygroups/{ENTRY_GROUP_ID}/entries/{ENTRY_ID}`.
Output only. The table name prefix of the shards. The name of any given shard is `[table_prefix]YYYYMMDD`. For example, for the `MyTable20180101` shard, the `table_prefix` is `MyTable`.
Output only. Total number of shards.
Output only. BigQuery resource name of the latest shard.
Fields specific for BigQuery routines.
Used in:
Paths of the imported libraries.
Describes a BigQuery table.
Used in:
Output only. The table source type.
Output only.
Table view specification. Populated only if the `table_source_type` is `BIGQUERY_VIEW`.
Specification of a BigQuery table. Populated only if the `table_source_type` is `BIGQUERY_TABLE`.
Business Context of the entry.
Used in:
Entry overview fields for rich text descriptions of entries.
Contact people for the entry.
Configuration related to the opt-in status for the UI switch to Dataplex.
Used in:
,Default value. The default UI is Dataplex.
The UI is Dataplex.
The UI is Data Catalog.
Specification that applies to Instance entries that are part of `CLOUD_BIGTABLE` system. (user_specified_type)
Used in:
The list of clusters for the Instance.
Spec that applies to clusters of an Instance of Cloud Bigtable.
Used in:
Name of the cluster.
Location of the cluster, typically a Cloud zone.
Type of the resource. For a cluster this would be "CLUSTER".
A link back to the parent resource, in this case Instance.
Specification that applies to all entries that are part of `CLOUD_BIGTABLE` system (user_specified_type)
Used in:
Display name of the Instance. This is user specified and different from the resource name.
Specification for the BigQuery connection to a Cloud SQL instance.
Used in:
Cloud SQL instance ID in the format of `project:location:instance`.
Database name.
Type of the Cloud SQL database.
Supported Cloud SQL database types.
Used in:
Unspecified database type.
Cloud SQL for PostgreSQL.
Cloud SQL for MySQL.
A column within a schema. Columns can be nested inside other columns.
Used in:
Required. Name of the column. Must be a UTF-8 string without dots (.). The maximum size is 64 bytes.
Required. Type of the column. Must be a UTF-8 string with the maximum size of 128 bytes.
Optional. Description of the column. Default value is an empty string. The description must be a UTF-8 string with the maximum size of 2000 bytes.
Optional. A column's mode indicates whether values in this column are required, nullable, or repeated. Only `NULLABLE`, `REQUIRED`, and `REPEATED` values are supported. Default mode is `NULLABLE`.
Optional. Default value for the column.
Optional. Ordinal position
Optional. Most important inclusion of this column.
Optional. Schema of sub-columns. A column can have zero or more sub-columns.
Information only applying for columns in Entries from a specific system.
Looker specific column info of this column.
Optional. The subtype of the RANGE, if the type of this field is RANGE. If the type is RANGE, this field is required. Possible values for the field element type of a RANGE include: * DATE * DATETIME * TIMESTAMP
Optional. Garbage collection policy for the column or column family. Applies to systems like Cloud Bigtable.
Represents the type of a field element.
Used in:
Required. The type of a field element. See [ColumnSchema.type][google.cloud.datacatalog.v1.ColumnSchema.type].
Specifies inclusion of the column in an index
Used in:
Unspecified.
Column not a part of an index.
Column Part of non unique index.
Column part of unique index.
Column part of the primary key.
Column info specific to Looker System.
Used in:
Looker specific column type of this column.
Column type in Looker.
Used in:
Unspecified.
Dimension.
Dimension group - parent for Dimension.
Filter.
Measure.
Parameter.
Common statistics on the entry's usage. They can be set on any system.
Used in:
View count in source system.
Contact people for the entry.
Used as response type in: DataCatalog.ModifyEntryContacts
Used as field type in:
,The list of contact people for the entry.
A contact person for the entry.
Used in:
Designation of the person, for example, Data Steward.
Email of the person in the format of `john.doe@xyz`, `<john.doe@xyz>`, or `John Doe<john.doe@xyz>`.
Cross-regional source used to import an existing taxonomy into a different region.
Used in:
Required. The resource name of the source taxonomy to import.
Physical location of an entry.
Used in:
Service that physically stores the data.
Full name of a resource as defined by the service. For example: `//bigquery.googleapis.com/projects/{PROJECT_ID}/locations/{LOCATION}/datasets/{DATASET_ID}/tables/{TABLE_ID}`
Output only. Data Catalog entry name, if applicable.
Detailed properties of the underlying storage.
Name of a service that stores the data.
Used in:
Default unknown service.
Google Cloud Storage service.
BigQuery service.
Specification that applies to a data source connection. Valid only for entries with the `DATA_SOURCE_CONNECTION` type. Only one of internal specs can be set at the time, and cannot be changed later.
Used in:
Output only. Fields specific to BigQuery connections.
Specification that applies to a table resource. Valid only for entries with the `TABLE` type.
Used in:
Type of this table.
Output only. Fields specific to a Dataplex table and present only in the Dataplex table entries.
Spec what applies to tables that are actually views. Not set for "real" tables.
Specification that applies to database view.
Used in:
Type of this view.
Definition of the view.
Name of a singular table this view reflects one to one.
SQL query used to generate this view.
Concrete type of the view.
Used in:
Default unknown view type.
Standard view.
Materialized view.
Type of the table.
Used in:
Default unknown table type.
Native table.
External table.
External table registered by Dataplex. Dataplex publishes data discovered from an asset into multiple other systems (BigQuery, DPMS) in form of tables. We call them "external tables". External tables are also synced into the Data Catalog. This message contains pointers to those external tables (fully qualified name, resource name et cetera) within the Data Catalog.
Used in:
Service in which the external table is registered.
Fully qualified name (FQN) of the external table.
Google Cloud resource name of the external table.
Name of the Data Catalog entry representing the external table.
Entry specyfication for a Dataplex fileset.
Used in:
Common Dataplex fields.
Common Dataplex fields.
Used in:
,Fully qualified resource name of an asset in Dataplex, to which the underlying data source (Cloud Storage bucket or BigQuery dataset) of the entity is attached.
Format of the data.
Compression format of the data, e.g., zip, gzip etc.
Project ID of the underlying Cloud Storage or BigQuery data. Note that this may not be the same project as the correspondingly Dataplex lake / zone / asset.
Entry specification for a Dataplex table.
Used in:
List of external tables registered by Dataplex in other systems based on the same underlying data. External tables allow to query this data in those systems.
Common Dataplex fields.
Indicates if the table schema is managed by the user or not.
Specification that applies to a dataset. Valid only for entries with the `DATASET` type.
Used in:
Fields specific to the source system.
Vertex AI Dataset specific fields
Wrapper for any item that can be contained in the dump.
Entry and its tags.
Entry metadata. A Data Catalog entry represents another resource in Google Cloud Platform (such as a BigQuery dataset or a Pub/Sub topic) or outside of it. You can use the `linked_resource` field in the entry resource to refer to the original resource ID of the source system. An entry resource contains resource details, for example, its schema. Additionally, you can attach flexible metadata to an entry in the form of a [Tag][google.cloud.datacatalog.v1.Tag].
Used as response type in: DataCatalog.CreateEntry, DataCatalog.GetEntry, DataCatalog.LookupEntry, DataCatalog.UpdateEntry
Used as field type in:
, , ,Output only. Identifier. The resource name of an entry in URL format. Note: The entry itself and its child resources might not be stored in the location specified in its name.
The resource this metadata entry refers to. For Google Cloud Platform resources, `linked_resource` is the [Full Resource Name] (https://cloud.google.com/apis/design/resource_names#full_resource_name). For example, the `linked_resource` for a table resource from BigQuery is: `//bigquery.googleapis.com/projects/{PROJECT_ID}/datasets/{DATASET_ID}/tables/{TABLE_ID}` Output only when the entry is one of the types in the `EntryType` enum. For entries with a `user_specified_type`, this field is optional and defaults to an empty string. The resource string must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), periods (.), colons (:), slashes (/), dashes (-), and hashes (#). The maximum size is 200 bytes when encoded in UTF-8.
[Fully Qualified Name (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names) of the resource. Set automatically for entries representing resources from synced systems. Settable only during creation, and read-only later. Can be used for search and lookup of the entries.
Required. Entry type.
The type of the entry. For details, see [`EntryType`](#entrytype).
Custom entry type that doesn't match any of the values allowed for input and listed in the `EntryType` enum. When creating an entry, first check the type values in the enum. If there are no appropriate types for the new entry, provide a custom value, for example, `my_special_type`. The `user_specified_type` string has the following limitations: * Is case insensitive. * Must begin with a letter or underscore. * Can only contain letters, numbers, and underscores. * Must be at least 1 character and at most 64 characters long.
The source system of the entry.
Output only. Indicates the entry's source system that Data Catalog integrates with, such as BigQuery, Pub/Sub, or Dataproc Metastore.
Indicates the entry's source system that Data Catalog doesn't automatically integrate with. The `user_specified_system` string has the following limitations: * Is case insensitive. * Must begin with a letter or underscore. * Can only contain letters, numbers, and underscores. * Must be at least 1 character and at most 64 characters long.
System specification. Can be used as a complement for `spec`, when some metadata is relevant for all entries existing within given system
Specification that applies to a relational database system. Only settable when `user_specified_system` is equal to `SQL_DATABASE`
Specification that applies to Looker sysstem. Only settable when `user_specified_system` is equal to `LOOKER`
Specification that applies to Cloud Bigtable system. Only settable when `integrated_system` is equal to `CLOUD_BIGTABLE`
Type specification.
Specification that applies to a Cloud Storage fileset. Valid only for entries with the `FILESET` type.
Output only. Specification that applies to a BigQuery table. Valid only for entries with the `TABLE` type.
Output only. Specification for a group of BigQuery tables with the `[prefix]YYYYMMDD` name pattern. For more information, see [Introduction to partitioned tables] (https://cloud.google.com/bigquery/docs/partitioned-tables#partitioning_versus_sharding).
Type- and system-specific information. Specifications for types contain fields common to all entries of a given type, and sub-specifications with fields specific to a given source system. When extending the API with new types and systems, use this field instead of the legacy `type_spec`.
Specification that applies to a table resource. Valid only for entries with the `TABLE` or `EXPLORE` type.
Specification that applies to a data source connection. Valid only for entries with the `DATA_SOURCE_CONNECTION` type.
Specification that applies to a user-defined function or procedure. Valid only for entries with the `ROUTINE` type.
Specification that applies to a dataset.
Specification that applies to a fileset resource. Valid only for entries with the `FILESET` type.
Specification that applies to a Service resource.
Model specification.
FeatureonlineStore spec for Vertex AI Feature Store.
Display name of an entry. The maximum size is 500 bytes when encoded in UTF-8. Default value is an empty string.
Entry description that can consist of several sentences or paragraphs that describe entry contents. The description must not contain Unicode non-characters as well as C0 and C1 control codes except tabs (HT), new lines (LF), carriage returns (CR), and page breaks (FF). The maximum size is 2000 bytes when encoded in UTF-8. Default value is an empty string.
Business Context of the entry. Not supported for BigQuery datasets
Schema of the entry. An entry might not have any schema attached to it.
Timestamps from the underlying resource, not from the Data Catalog entry. Output only when the entry has a system listed in the `IntegratedSystem` enum. For entries with `user_specified_system`, this field is optional and defaults to an empty timestamp.
Resource usage statistics.
Cloud labels attached to the entry. In Data Catalog, you can create and modify labels attached only to custom entries. Synced entries have unmodifiable labels that come from the source system.
Output only. Physical location of the entry.
Output only. Additional information related to the entry. Private to the current user.
Entry group metadata. An `EntryGroup` resource represents a logical grouping of zero or more Data Catalog [Entry][google.cloud.datacatalog.v1.Entry] resources.
Used as response type in: DataCatalog.CreateEntryGroup, DataCatalog.GetEntryGroup, DataCatalog.UpdateEntryGroup
Used as field type in:
, ,Identifier. The resource name of the entry group in URL format. Note: The entry group itself and its child resources might not be stored in the location specified in its name.
A short name to identify the entry group, for example, "analytics data - jan 2011". Default value is an empty string.
Entry group description. Can consist of several sentences or paragraphs that describe the entry group contents. Default value is an empty string.
Output only. Timestamps of the entry group. Default value is empty.
Optional. When set to [true], it means DataCatalog EntryGroup was transferred to Dataplex Catalog Service. It makes EntryGroup and its Entries to be read-only in DataCatalog. However, new Tags on EntryGroup and its Entries can be created. After setting the flag to [true] it cannot be unset.
Entry overview fields for rich text descriptions of entries.
Used as response type in: DataCatalog.ModifyEntryOverview
Used as field type in:
,Entry overview with support for rich text. The overview must only contain Unicode characters, and should be formatted using HTML. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
Metadata automatically ingested from Google Cloud resources like BigQuery tables or Pub/Sub topics always uses enum values from `EntryType` as the type of entry. Other sources of metadata like Hive or Oracle databases can identify the type by either using one of the enum values from `EntryType` (for example, `FILESET` for a Cloud Storage fileset) or specifying a custom value using the [`Entry`](#resource:-entry) field `user_specified_type`. For more information, see [Surface files from Cloud Storage with fileset entries](/data-catalog/docs/how-to/filesets) or [Create custom entries for your data sources](/data-catalog/docs/how-to/custom-entries).
Used in:
Default unknown type.
The entry type that has a GoogleSQL schema, including logical views.
The type of models. For more information, see [Supported models in BigQuery ML](/bigquery/docs/bqml-introduction#supported_models).
An entry type for streaming entries. For example, a Pub/Sub topic.
An entry type for a set of files or objects. For example, a Cloud Storage fileset.
A group of servers that work together. For example, a Kafka cluster.
A database.
Connection to a data source. For example, a BigQuery connection.
Routine, for example, a BigQuery routine.
A Dataplex lake.
A Dataplex zone.
A service, for example, a Dataproc Metastore service.
Schema within a relational database.
A Dashboard, for example from Looker.
A Looker Explore. For more information, see [Looker Explore API] (https://developers.looker.com/api/explorer/4.0/methods/LookmlModel/lookml_model_explore).
A Looker Look. For more information, see [Looker Look API] (https://developers.looker.com/api/explorer/4.0/methods/Look).
Feature Online Store resource in Vertex AI Feature Store.
Feature View resource in Vertex AI Feature Store.
Feature Group resource in Vertex AI Feature Store.
Detail description of the source information of a Vertex Feature Online Store.
Used in:
Output only. Type of underlying storage for the FeatureOnlineStore.
Type of underlying storage type.
Used in:
Should not be used.
Underlsying storgae is Bigtable.
Underlying is optimized online server (Lightning).
Used in:
Required.
Primitive types, such as string, boolean, etc.
An enum type.
Used in:
The set of allowed values for this enum. This set must not be empty and can include up to 100 allowed values. The display names of the values in this set must not be empty and must be case-insensitively unique within this set. The order of items in this set is preserved. This field can be used to create, remove, and reorder enum values. To rename enum values, use the `RenameTagTemplateFieldEnumValue` method.
Used in:
Required. The display name of the enum value. Must not be an empty string. The name must contain only Unicode letters, numbers (0-9), underscores (_), dashes (-), spaces ( ), and can't start or end with spaces. The maximum length is 200 characters.
Used in:
The default invalid value for a type.
A double precision number.
An UTF-8 string.
A boolean value.
A timestamp.
A Richtext description.
Specification that applies to a fileset. Valid only for entries with the 'FILESET' type.
Used in:
Fields specific to a Dataplex fileset and present only in the Dataplex fileset entries.
Specification of a single file in Cloud Storage.
Used in:
Required. Full file path. Example: `gs://bucket_name/a/b.txt`.
Output only. Creation, modification, and expiration timestamps of a Cloud Storage file.
Output only. File size in bytes.
Describes a Cloud Storage fileset entry.
Used in:
Required. Patterns to identify a set of files in Google Cloud Storage. For more information, see [Wildcard Names] (https://cloud.google.com/storage/docs/wildcards). Note: Currently, bucket wildcards are not supported. Examples of valid `file_patterns`: * `gs://bucket_name/dir/*`: matches all files in `bucket_name/dir` directory * `gs://bucket_name/dir/**`: matches all files in `bucket_name/dir` and all subdirectories * `gs://bucket_name/file*`: matches files prefixed by `file` in `bucket_name` * `gs://bucket_name/??.txt`: matches files with two characters followed by `.txt` in `bucket_name` * `gs://bucket_name/[aeiou].txt`: matches files that contain a single vowel character followed by `.txt` in `bucket_name` * `gs://bucket_name/[a-m].txt`: matches files that contain `a`, `b`, ... or `m` followed by `.txt` in `bucket_name` * `gs://bucket_name/a/*/b`: matches all files in `bucket_name` that match the `a/*/b` pattern, such as `a/c/b`, `a/d/b` * `gs://another_bucket/a.txt`: matches `gs://another_bucket/a.txt` You can combine wildcards to match complex sets of files, for example: `gs://bucket_name/[a-m]??.j*g`
Output only. Sample files contained in this fileset, not all files contained in this fileset are represented here.
Metadata message for [long-running operation][google.longrunning.Operation] returned by the [ImportEntries][google.cloud.datacatalog.v1.DataCatalog.ImportEntries].
State of the import operation.
Partial errors that are encountered during the ImportEntries operation. There is no guarantee that all the encountered errors are reported. However, if no errors are reported, it means that no errors were encountered.
Enum holding possible states of the import operation.
Used in:
Default value. This value is unused.
The dump with entries has been queued for import.
The import of entries is in progress.
The import of entries has been finished.
The import of entries has been abandoned in favor of a newer request.
Response message for [long-running operation][google.longrunning.Operation] returned by the [ImportEntries][google.cloud.datacatalog.v1.DataCatalog.ImportEntries].
Cumulative number of entries created and entries updated as a result of import operation.
Number of entries deleted as a result of import operation.
Inline source containing taxonomies to import.
Used in:
Required. Taxonomies to import.
This enum lists all the systems that Data Catalog integrates with.
Used in:
, ,Default unknown system.
BigQuery.
Cloud Pub/Sub.
Dataproc Metastore.
Dataplex.
Cloud Spanner
Cloud Bigtable
Cloud Sql
Looker
Vertex AI
Specification that applies to entries that are part `LOOKER` system (user_specified_type)
Used in:
ID of the parent Looker Instance. Empty if it does not exist. Example value: `someinstance.looker.com`
Name of the parent Looker Instance. Empty if it does not exist.
ID of the parent Model. Empty if it does not exist.
Name of the parent Model. Empty if it does not exist.
ID of the parent View. Empty if it does not exist.
Name of the parent View. Empty if it does not exist.
This enum describes all the systems that manage Taxonomy and PolicyTag resources in DataCatalog.
Used in:
Default value
Dataplex.
Other
The configuration related to the migration to Dataplex applied to an organization or project. It is the response message for [SetConfig][google.cloud.datacatalog.v1.DataCatalog.SetConfig] and [RetrieveEffectiveConfig][google.cloud.datacatalog.v1.DataCatalog.RetrieveEffectiveConfig].
Used as response type in: DataCatalog.RetrieveEffectiveConfig, DataCatalog.SetConfig
Used as field type in:
Opt-in status for the migration of Tag Templates to Dataplex.
Opt-in status for the UI switch to Dataplex.
The time when the Tag Template migration was enabled. If the Tag Template migration is not enabled, this field is not set.
Specification that applies to a model. Valid only for entries with the `MODEL` type.
Used in:
System spec
Specification for vertex model resources.
Entry metadata relevant only to the user and private to them.
Used in:
True if the entry is starred by the user; false otherwise.
Set if the entry is starred; unset otherwise.
Native schema used by a resource represented as an entry. Used by query engines for deserializing and parsing source data.
Used in:
Schema in Avro JSON format.
Schema in Thrift format.
Schema in protocol buffer format.
Marks a Parquet-encoded data source.
Marks an ORC-encoded data source.
Marks a CSV-encoded data source.
Schema in Avro JSON format.
Used in:
JSON source of the Avro schema.
Marks a CSV-encoded data source.
Used in:
(message has no fields)
Marks an ORC-encoded data source.
Used in:
(message has no fields)
Marks a Parquet-encoded data source.
Used in:
(message has no fields)
Schema in protocol buffer format.
Used in:
Protocol buffer source of the schema.
Schema in Thrift format.
Used in:
Thrift IDL source of the schema.
Denotes one policy tag in a taxonomy, for example, SSN. Policy tags can be defined in a hierarchy. For example: ``` + Geolocation + LatLong + City + ZipCode ``` Where the "Geolocation" policy tag contains three children.
Used as response type in: PolicyTagManager.CreatePolicyTag, PolicyTagManager.GetPolicyTag, PolicyTagManager.UpdatePolicyTag
Used as field type in:
, ,Identifier. Resource name of this policy tag in the URL format. The policy tag manager generates unique taxonomy IDs and policy tag IDs.
Required. User-defined name of this policy tag. The name can't start or end with spaces and must be unique within the parent taxonomy, contain only Unicode letters, numbers, underscores, dashes and spaces, and be at most 200 bytes long when encoded in UTF-8.
Description of this policy tag. If not set, defaults to empty. The description must contain only Unicode characters, tabs, newlines, carriage returns and page breaks, and be at most 2000 bytes long when encoded in UTF-8.
Resource name of this policy tag's parent policy tag. If empty, this is a top level tag. If not set, defaults to an empty string. For example, for the "LatLong" policy tag in the example above, this field contains the resource name of the "Geolocation" policy tag, and, for "Geolocation", this field is empty.
Output only. Resource names of child policy tags of this policy tag.
[Long-running operation][google.longrunning.Operation] metadata message returned by the [ReconcileTags][google.cloud.datacatalog.v1.DataCatalog.ReconcileTags].
State of the reconciliation operation.
Maps the name of each tagged column (or empty string for a sole entry) to tagging operation [status][google.rpc.Status].
Enum holding possible states of the reconciliation operation.
Used in:
Default value. This value is unused.
The reconciliation has been queued and awaits for execution.
The reconciliation is in progress.
The reconciliation has been finished.
[Long-running operation][google.longrunning.Operation] response message returned by [ReconcileTags][google.cloud.datacatalog.v1.DataCatalog.ReconcileTags].
Number of tags created in the request.
Number of tags updated in the request.
Number of tags deleted in the request.
Specification that applies to a routine. Valid only for entries with the `ROUTINE` type.
Used in:
The type of the routine.
The language the routine is written in. The exact value depends on the source system. For BigQuery routines, possible values are: * `SQL` * `JAVASCRIPT`
Arguments of the routine.
Return type of the argument. The exact value depends on the source system and the language.
The body of the routine.
Contains fields specific to the source system.
Fields specific for BigQuery routines.
Input or output argument of a function or stored procedure.
Used in:
The name of the argument. A return argument of a function might not have a name.
Specifies whether the argument is input or output.
Type of the argument. The exact value depends on the source system and the language.
The input or output mode of the argument.
Used in:
Unspecified mode.
The argument is input-only.
The argument is output-only.
The argument is both an input and an output.
The fine-grained type of the routine.
Used in:
Unspecified type.
Non-builtin permanent scalar function.
Stored procedure.
Represents a schema, for example, a BigQuery, GoogleSQL, or Avro schema.
Used in:
The unified GoogleSQL-like schema of columns. The overall maximum number of columns and nested columns is 10,000. The maximum nested depth is 15 levels.
The criteria that select the subspace used for query matching.
Used in:
The list of organization IDs to search within. To find your organization ID, follow the steps from [Creating and managing organizations] (/resource-manager/docs/creating-managing-organization).
The list of project IDs to search within. For more information on the distinction between project names, IDs, and numbers, see [Projects](/docs/overview/#projects).
If `true`, include Google Cloud public datasets in search results. By default, they are excluded. See [Google Cloud Public Datasets](/public-datasets) for more information.
Optional. The list of locations to search within. If empty, all locations are searched. Returns an error if any location in the list isn't one of the [Supported regions](https://cloud.google.com/data-catalog/docs/concepts/regions#supported_regions). If a location is unreachable, its name is returned in the `SearchCatalogResponse.unreachable` field. To get additional information on the error, repeat the search request and set the location name as the value of this parameter.
Optional. If `true`, search only among starred entries. By default, all results are returned, starred or not.
Optional. This field is deprecated. The search mechanism for public and private tag templates is the same.
Result in the response to a search request. Each result captures details of one entry that matches the search.
Used in:
Type of the search result. You can use this field to determine which get method to call to fetch the full resource.
Sub-type of the search result. A dot-delimited full type of the resource. The same type you specify in the `type` search predicate. Examples: `entry.table`, `entry.dataStream`, `tagTemplate`.
The relative name of the resource in URL format. Examples: * `projects/{PROJECT_ID}/locations/{LOCATION_ID}/entryGroups/{ENTRY_GROUP_ID}/entries/{ENTRY_ID}` * `projects/{PROJECT_ID}/tagTemplates/{TAG_TEMPLATE_ID}`
The full name of the Google Cloud resource the entry belongs to. For more information, see [Full Resource Name] (/apis/design/resource_names#full_resource_name). Example: `//bigquery.googleapis.com/projects/PROJECT_ID/datasets/DATASET_ID/tables/TABLE_ID`
The last modification timestamp of the entry in the source system.
The source system of the entry. Applicable only when the `search_result_type` is `ENTRY`.
Output only. The source system that Data Catalog automatically integrates with, such as BigQuery, Cloud Pub/Sub, or Dataproc Metastore.
Custom source system that you can manually integrate Data Catalog with.
Fully qualified name (FQN) of the resource. FQNs take two forms: * For non-regionalized resources: `{SYSTEM}:{PROJECT}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` * For regionalized resources: `{SYSTEM}:{PROJECT}.{LOCATION_ID}.{PATH_TO_RESOURCE_SEPARATED_WITH_DOTS}` Example for a DPMS table: `dataproc_metastore:PROJECT_ID.LOCATION_ID.INSTANCE_ID.DATABASE_ID.TABLE_ID`
The display name of the result.
Entry description that can consist of several sentences or paragraphs that describe entry contents.
The resource types that can be returned in search results.
Used in:
Default unknown type.
An [Entry][google.cloud.datacatalog.v1.Entry].
A [TagTemplate][google.cloud.datacatalog.v1.TagTemplate].
An [EntryGroup][google.cloud.datacatalog.v1.EntryGroup].
A nested protocol buffer that represents a policy tag and all its descendants.
Used in:
Resource name of the policy tag. This field is ignored when calling `ImportTaxonomies`.
Required. Display name of the policy tag. At most 200 bytes when encoded in UTF-8.
Description of the serialized policy tag. At most 2000 bytes when encoded in UTF-8. If not set, defaults to an empty description.
Children of the policy tag, if any.
A nested protocol buffer that represents a taxonomy and the hierarchy of its policy tags. Used for taxonomy replacement, import, and export.
Used in:
, ,Required. Display name of the taxonomy. At most 200 bytes when encoded in UTF-8.
Description of the serialized taxonomy. At most 2000 bytes when encoded in UTF-8. If not set, defaults to an empty description.
Top level policy tags associated with the taxonomy, if any.
A list of policy types that are activated per taxonomy.
Specification that applies to a Service resource. Valid only for entries with the `SERVICE` type.
Used in:
System spec
Specification that applies to Instance entries of `CLOUD_BIGTABLE` system.
Specification that applies to entries that are part `SQL_DATABASE` system (user_specified_type)
Used in:
SQL Database Engine. enum SqlEngine { UNDEFINED = 0; MY_SQL = 1; POSTGRE_SQL = 2; SQL_SERVER = 3; } Engine of the enclosing database instance.
Version of the database engine.
Host of the SQL database enum InstanceHost { UNDEFINED = 0; SELF_HOSTED = 1; CLOUD_SQL = 2; AMAZON_RDS = 3; AZURE_SQL = 4; } Host of the enclousing database instance.
Details the properties of the underlying storage.
Used in:
Patterns to identify a set of files for this fileset. Examples of a valid `file_pattern`: * `gs://bucket_name/dir/*`: matches all files in the `bucket_name/dir` directory * `gs://bucket_name/dir/**`: matches all files in the `bucket_name/dir` and all subdirectories recursively * `gs://bucket_name/file*`: matches files prefixed by `file` in `bucket_name` * `gs://bucket_name/??.txt`: matches files with two characters followed by `.txt` in `bucket_name` * `gs://bucket_name/[aeiou].txt`: matches files that contain a single vowel character followed by `.txt` in `bucket_name` * `gs://bucket_name/[a-m].txt`: matches files that contain `a`, `b`, ... or `m` followed by `.txt` in `bucket_name` * `gs://bucket_name/a/*/b`: matches all files in `bucket_name` that match the `a/*/b` pattern, such as `a/c/b`, `a/d/b` * `gs://another_bucket/a.txt`: matches `gs://another_bucket/a.txt`
File type in MIME format, for example, `text/plain`.
Timestamps associated with this resource in a particular system.
Used in:
, , ,Creation timestamp of the resource within the given system.
Timestamp of the last modification of the resource or its metadata within a given system. Note: Depending on the source system, not every modification updates this timestamp. For example, BigQuery timestamps every metadata modification but not data or permission changes.
Output only. Expiration timestamp of the resource within the given system. Currently only applicable to BigQuery resources.
Table source type.
Used in:
Default unknown type.
Table view.
BigQuery native table.
BigQuery materialized view.
Normal BigQuery table specification.
Used in:
Output only. If the table is date-sharded, that is, it matches the `[prefix]YYYYMMDD` name pattern, this field is the Data Catalog resource name of the date-sharded grouped entry. For example: `projects/{PROJECT_ID}/locations/{LOCATION}/entrygroups/{ENTRY_GROUP_ID}/entries/{ENTRY_ID}`. Otherwise, `grouped_entry` is empty.
Tags contain custom metadata and are attached to Data Catalog resources. Tags conform with the specification of their tag template. See [Data Catalog IAM](https://cloud.google.com/data-catalog/docs/concepts/iam) for information on the permissions needed to create or view tags.
Used as response type in: DataCatalog.CreateTag, DataCatalog.UpdateTag
Used as field type in:
, , , ,Identifier. The resource name of the tag in URL format where tag ID is a system-generated identifier. Note: The tag itself might not be stored in the location specified in its name.
Required. The resource name of the tag template this tag uses. Example: `projects/{PROJECT_ID}/locations/{LOCATION}/tagTemplates/{TAG_TEMPLATE_ID}` This field cannot be modified after creation.
Output only. The display name of the tag template.
The scope within the parent resource that this tag is attached to. If not provided, the tag is attached to the parent resource itself. Deleting the scope from the parent resource deletes all tags attached to that scope. These fields cannot be updated after creation.
Resources like entry can have schemas associated with them. This scope allows you to attach tags to an individual column based on that schema. To attach a tag to a nested column, separate column names with a dot (`.`). Example: `column.nested_column`.
Required. Maps the ID of a tag field to its value and additional information about that field. Tag template defines valid field IDs. A tag must have at least 1 field and at most 500 fields.
Output only. Denotes the transfer status of the Tag Template.
Contains the value and additional information on a field within a [Tag][google.cloud.datacatalog.v1.Tag].
Used in:
Output only. The display name of this field.
Required. The value of this field.
The value of a tag field with a double type.
The value of a tag field with a string type. The maximum length is 2000 UTF-8 characters.
The value of a tag field with a boolean type.
The value of a tag field with a timestamp type.
The value of a tag field with an enum type. This value must be one of the allowed values listed in this enum.
The value of a tag field with a rich text type. The maximum length is 10 MiB as this value holds HTML descriptions including encoded images. The maximum length of the text without images is 100 KiB.
Output only. The order of this field with respect to other fields in this tag. Can be set by [Tag][google.cloud.datacatalog.v1.TagTemplateField.order]. For example, a higher value can indicate a more important field. The value can be negative. Multiple fields can have the same order, and field orders within a tag don't have to be sequential.
An enum value.
Used in:
The display name of the enum value.
A tag template defines a tag that can have one or more typed fields. The template is used to create tags that are attached to Google Cloud resources. [Tag template roles] (https://cloud.google.com/iam/docs/understanding-roles#data-catalog-roles) provide permissions to create, edit, and use the template. For example, see the [TagTemplate User] (https://cloud.google.com/data-catalog/docs/how-to/template-user) role that includes a permission to use the tag template to tag resources.
Used as response type in: DataCatalog.CreateTagTemplate, DataCatalog.GetTagTemplate, DataCatalog.UpdateTagTemplate
Used as field type in:
,Identifier. The resource name of the tag template in URL format. Note: The tag template itself and its child resources might not be stored in the location specified in its name.
Display name for this template. Defaults to an empty string. The name must contain only Unicode letters, numbers (0-9), underscores (_), dashes (-), spaces ( ), and can't start or end with spaces. The maximum length is 200 characters.
Indicates whether tags created with this template are public. Public tags do not require tag template access to appear in [ListTags][google.cloud.datacatalog.v1.DataCatalog.ListTags] API response. Additionally, you can search for a public tag by value with a simple search query in addition to using a ``tag:`` predicate.
Required. Map of tag template field IDs to the settings for the field. This map is an exhaustive list of the allowed fields. The map must contain at least one field and at most 500 fields. The keys to this map are tag template field IDs. The IDs have the following limitations: * Can contain uppercase and lowercase letters, numbers (0-9) and underscores (_). * Must be at least 1 character and at most 64 characters long. * Must start with a letter or underscore.
Optional. Transfer status of the TagTemplate
This enum describes TagTemplate transfer status to Dataplex service.
Used in:
,Default value. TagTemplate and its tags are only visible and editable in DataCatalog.
TagTemplate and its tags are auto-copied to Dataplex service. Visible in both services. Editable in DataCatalog, read-only in Dataplex. Deprecated: Individual TagTemplate migration is deprecated in favor of organization or project wide TagTemplate migration opt-in.
TagTemplate and its tags are auto-copied to Dataplex service. Visible in both services. Editable in Dataplex, read-only in DataCatalog.
The template for an individual field within a tag template.
Used as response type in: DataCatalog.CreateTagTemplateField, DataCatalog.RenameTagTemplateField, DataCatalog.RenameTagTemplateFieldEnumValue, DataCatalog.UpdateTagTemplateField
Used as field type in:
, ,Identifier. The resource name of the tag template field in URL format. Example: `projects/{PROJECT_ID}/locations/{LOCATION}/tagTemplates/{TAG_TEMPLATE}/fields/{FIELD}` Note: The tag template field itself might not be stored in the location specified in its name. The name must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_), and must start with a letter or underscore. The maximum length is 64 characters.
The display name for this field. Defaults to an empty string. The name must contain only Unicode letters, numbers (0-9), underscores (_), dashes (-), spaces ( ), and can't start or end with spaces. The maximum length is 200 characters.
Required. The type of value this tag field can contain.
If true, this field is required. Defaults to false.
The description for this field. Defaults to an empty string.
The order of this field with respect to other fields in this tag template. For example, a higher value can indicate a more important field. The value can be negative. Multiple fields can have the same order and field orders within a tag don't have to be sequential.
Configuration related to the opt-in status for the migration of TagTemplates to Dataplex.
Used in:
,Default value. Migration of Tag Templates from Data Catalog to Dataplex is not performed.
Migration of Tag Templates from Data Catalog to Dataplex is enabled.
Migration of Tag Templates from Data Catalog to Dataplex is disabled.
Wrapper containing Entry and information about Tags that should and should not be attached to it.
Used in:
Required. Entry to be ingested.
Non-encrypted Data Catalog v1 Entry.
Optional. Tags that should be ingested into the Data Catalog. Caller should populate template name, column and fields.
Optional. Tags that should be deleted from the Data Catalog. Caller should populate template name and column only.
A taxonomy is a collection of hierarchical policy tags that classify data along a common axis. For example, a "data sensitivity" taxonomy might contain the following policy tags: ``` + PII + Account number + Age + SSN + Zipcode + Financials + Revenue ``` A "data origin" taxonomy might contain the following policy tags: ``` + User data + Employee data + Partner data + Public data ```
Used as response type in: PolicyTagManager.CreateTaxonomy, PolicyTagManager.GetTaxonomy, PolicyTagManager.UpdateTaxonomy, PolicyTagManagerSerialization.ReplaceTaxonomy
Used as field type in:
, , ,Identifier. Resource name of this taxonomy in URL format. Note: Policy tag manager generates unique taxonomy IDs.
Required. User-defined name of this taxonomy. The name can't start or end with spaces, must contain only Unicode letters, numbers, underscores, dashes, and spaces, and be at most 200 bytes long when encoded in UTF-8. The taxonomy display name must be unique within an organization.
Optional. Description of this taxonomy. If not set, defaults to empty. The description must contain only Unicode characters, tabs, newlines, carriage returns, and page breaks, and be at most 2000 bytes long when encoded in UTF-8.
Output only. Number of policy tags in this taxonomy.
Output only. Creation and modification timestamps of this taxonomy.
Optional. A list of policy types that are activated for this taxonomy. If not set, defaults to an empty list.
Output only. Identity of the service which owns the Taxonomy. This field is only populated when the taxonomy is created by a Google Cloud service. Currently only 'DATAPLEX' is supported.
Defines policy types where the policy tags can be used for.
Used in:
,Unspecified policy type.
Fine-grained access control policy that enables access control on tagged sub-resources.
The source system of the Taxonomy.
Used in:
The Google Cloud service name.
The service agent for the service.
The set of all usage signals that Data Catalog stores. Note: Usually, these signals are updated daily. In rare cases, an update may fail but will be performed again on the next day.
Used in:
The end timestamp of the duration of usage statistics.
Output only. BigQuery usage statistics over each of the predefined time ranges. Supported time ranges are `{"24H", "7D", "30D"}`.
Common usage statistics over each of the predefined time ranges. Supported time ranges are `{"24H", "7D", "30D", "Lifetime"}`.
Favorite count in the source system.
Detailed statistics on the entry's usage. Usage statistics have the following limitations: - Only BigQuery tables have them. - They only include BigQuery query jobs. - They might be underestimated because wildcard table references are not yet counted. For more information, see [Querying multiple tables using a wildcard table] (https://cloud.google.com/bigquery/docs/querying-wildcard-tables)
Used in:
The number of successful uses of the underlying entry.
The number of failed attempts to use the underlying entry.
The number of cancelled attempts to use the underlying entry.
Total time spent only on successful uses, in milliseconds.
Specification for vertex dataset resources.
Used in:
The number of DataItems in this Dataset. Only apply for non-structured Dataset.
Type of the dataset.
Type of data stored in the dataset.
Used in:
Should not be used.
Structured data dataset.
Image dataset which supports ImageClassification, ImageObjectDetection and ImageSegmentation problems.
Document dataset which supports TextClassification, TextExtraction and TextSentiment problems.
Video dataset which supports VideoClassification, VideoObjectTracking and VideoActionRecognition problems.
Conversation dataset which supports conversation problems.
TimeSeries dataset.
Document dataset which supports DocumentAnnotation problems.
TextToSpeech dataset which supports TextToSpeech problems.
Translation dataset which supports Translation problems.
Store Vision dataset which is used for HITL integration.
Enterprise Knowledge Graph dataset which is used for HITL labeling integration.
Text prompt dataset which supports Large Language Models.
Detail description of the source information of a Vertex model.
Used in:
Type of the model source.
If this Model is copy of another Model. If true then [source_type][google.cloud.datacatalog.v1.VertexModelSourceInfo.source_type] pertains to the original.
Source of the model.
Used in:
Should not be used.
The Model is uploaded by automl training pipeline.
The Model is uploaded by user or custom training pipeline.
The Model is registered and sync'ed from BigQuery ML.
The Model is saved or tuned from Model Garden.
The Model is saved or tuned from Genie.
The Model is uploaded by text embedding finetuning pipeline.
The Model is saved or tuned from Marketplace.
Specification for vertex model resources.
Used in:
The version ID of the model.
User provided version aliases so that a model version can be referenced via alias
The description of this version.
Source of a Vertex model.
URI of the Docker image to be used as the custom container for serving predictions.
Table view specification.
Used in:
Output only. The query that defines the table view.