From dba582228c88ae4c930e6a8603ce191907d0618c Mon Sep 17 00:00:00 2001 From: yoshi-code-bot <70984784+yoshi-code-bot@users.noreply.github.com> Date: Mon, 4 Nov 2024 23:32:20 -0800 Subject: [PATCH] chore: Update discovery artifacts (#2512) ## Deleted keys were detected in the following stable discovery artifacts: beyondcorp v1 https://togithub.com/googleapis/google-api-python-client/commit/a7947120205705a0d4d10de082e7f46fc10a6edf developerconnect v1 https://togithub.com/googleapis/google-api-python-client/commit/d200c2ce1d4c280a5d214f74d301c7b3dd502450 ## Deleted keys were detected in the following pre-stable discovery artifacts: analyticshub v1beta1 https://togithub.com/googleapis/google-api-python-client/commit/2e8f827879892114930b2f973b1bc42d7e9dced1 ## Discovery Artifact Change Summary: feat(accesscontextmanager): update the api https://togithub.com/googleapis/google-api-python-client/commit/f2e314b119cedd5d66b0032d60e70c4a8daf6bb9 feat(admin): update the api https://togithub.com/googleapis/google-api-python-client/commit/5bda041fd412716b5684ee98fd1798b8c4b7b29c feat(aiplatform): update the api https://togithub.com/googleapis/google-api-python-client/commit/477d209fc46d85f7ac5384faef9f9537d85c422d feat(analyticshub): update the api https://togithub.com/googleapis/google-api-python-client/commit/2e8f827879892114930b2f973b1bc42d7e9dced1 feat(beyondcorp): update the api https://togithub.com/googleapis/google-api-python-client/commit/a7947120205705a0d4d10de082e7f46fc10a6edf feat(clouddeploy): update the api https://togithub.com/googleapis/google-api-python-client/commit/9f46ad86f2ca66f78325b30b897690435f1196fc feat(compute): update the api https://togithub.com/googleapis/google-api-python-client/commit/396f9dca70263bca2c0c2959ed333192c4785120 feat(container): update the api https://togithub.com/googleapis/google-api-python-client/commit/d35e92ef9e86dc1d494299a942f9e45aad6638a9 feat(dataflow): update the api https://togithub.com/googleapis/google-api-python-client/commit/6768b1f50b67aebb3ed19251101e545ce8b457d2 feat(datastream): update the api https://togithub.com/googleapis/google-api-python-client/commit/cdea035fc31dbdf929e9b6163417a9ece6b4dc66 feat(developerconnect): update the api https://togithub.com/googleapis/google-api-python-client/commit/d200c2ce1d4c280a5d214f74d301c7b3dd502450 feat(dialogflow): update the api https://togithub.com/googleapis/google-api-python-client/commit/d805cf2f5a85ffdf3c0e1784c243dc41b6a0b8f3 feat(discoveryengine): update the api https://togithub.com/googleapis/google-api-python-client/commit/cb59b84af6e4422494c730775ff8ed2a67fc9cdf feat(firebaseappdistribution): update the api https://togithub.com/googleapis/google-api-python-client/commit/bb87986fb57af4fa7d1dc62290667617a61ea973 feat(gkehub): update the api https://togithub.com/googleapis/google-api-python-client/commit/c4ce0d990967cc3de53ccb3870950d2e12bcaf54 feat(gkeonprem): update the api https://togithub.com/googleapis/google-api-python-client/commit/cf051473a2096c9be43355b622a65777011c4080 feat(migrationcenter): update the api https://togithub.com/googleapis/google-api-python-client/commit/46b6cf4d16e9db27e2b3cf486b80daa6af0c1000 feat(networkmanagement): update the api https://togithub.com/googleapis/google-api-python-client/commit/d5e3d41c95cd43d1ac923f3e05a8f55eab9ef59f feat(serviceusage): update the api https://togithub.com/googleapis/google-api-python-client/commit/5f01e669132826613b2329369f79d246b3fbc4f8 feat(tpu): update the api https://togithub.com/googleapis/google-api-python-client/commit/0b9b48260f17a3b5675cbd88dc3dc0c63dbcd400 fix(translate): update the api https://togithub.com/googleapis/google-api-python-client/commit/b61649bddc2ddece3f2eda06ebff98a0a4f16b7d feat(youtube): update the api https://togithub.com/googleapis/google-api-python-client/commit/ef091d560f62d9f9bcbd629d355125fd91362a57 --- ...1.organizations.gcpUserAccessBindings.html | 38 +- .../admin_directory_v1.chromeosdevices.html | 24 + docs/dyn/aiplatform_v1.endpoints.html | 6 +- docs/dyn/aiplatform_v1.projects.html | 69 ++ ...tform_v1.projects.locations.endpoints.html | 99 +- ...jects.locations.notebookExecutionJobs.html | 72 ++ ....projects.locations.publishers.models.html | 99 +- ...tform_v1.projects.locations.schedules.html | 144 +++ ...form_v1.projects.locations.tuningJobs.html | 10 +- docs/dyn/aiplatform_v1.publishers.models.html | 6 +- docs/dyn/aiplatform_v1beta1.endpoints.html | 6 +- ...ta1.projects.locations.cachedContents.html | 12 +- ...cts.locations.deploymentResourcePools.html | 3 + ..._v1beta1.projects.locations.endpoints.html | 123 ++- ...v1beta1.projects.locations.extensions.html | 10 +- ...ps.featureMonitors.featureMonitorJobs.html | 346 +++++++ ...cations.featureGroups.featureMonitors.html | 299 ++++++ ...ects.locations.featureGroups.features.html | 67 +- ...eta1.projects.locations.featureGroups.html | 5 + ...ns.featurestores.entityTypes.features.html | 79 +- ...eta1.projects.locations.featurestores.html | 12 + ...jects.locations.notebookExecutionJobs.html | 72 ++ ....projects.locations.publishers.models.html | 99 +- ...1.projects.locations.reasoningEngines.html | 1 + ..._v1beta1.projects.locations.schedules.html | 144 +++ ...v1beta1.projects.locations.tuningJobs.html | 10 +- .../aiplatform_v1beta1.publishers.models.html | 6 +- ...projects.locations.clusters.instances.html | 10 +- ...projects.locations.clusters.instances.html | 10 +- ...projects.locations.clusters.instances.html | 10 +- ...b_v1.projects.locations.dataExchanges.html | 12 + ...oidmanagement_v1.enterprises.policies.html | 8 +- ....projects.locations.managementServers.html | 6 +- docs/dyn/beyondcorp_v1.html | 5 - ...1.projects.locations.securityGateways.html | 88 ++ docs/dyn/chat_v1.spaces.html | 6 +- docs/dyn/chat_v1.spaces.members.html | 54 +- docs/dyn/chat_v1.spaces.spaceEvents.html | 74 +- ...ions.deliveryPipelines.automationRuns.html | 118 ++- ...cations.deliveryPipelines.automations.html | 208 +++++ ...tions_v2.projects.locations.functions.html | 8 +- ..._v2alpha.projects.locations.functions.html | 8 +- ...s_v2beta.projects.locations.functions.html | 8 +- ...er_v1.projects.locations.environments.html | 48 +- ....environments.userWorkloadsConfigMaps.html | 12 +- ...ons.environments.userWorkloadsSecrets.html | 12 +- ...beta1.projects.locations.environments.html | 48 +- ....environments.userWorkloadsConfigMaps.html | 12 +- ...ons.environments.userWorkloadsSecrets.html | 12 +- docs/dyn/compute_v1.backendBuckets.html | 10 +- docs/dyn/compute_v1.backendServices.html | 14 +- docs/dyn/compute_v1.instances.html | 2 +- docs/dyn/compute_v1.networks.html | 12 + .../dyn/compute_v1.regionBackendServices.html | 12 +- docs/dyn/compute_v1.routers.html | 9 + docs/dyn/compute_v1.routes.html | 9 + docs/dyn/compute_v1.subnetworks.html | 12 +- ...r_v1beta1.projects.locations.clusters.html | 18 + ...ainer_v1beta1.projects.zones.clusters.html | 18 + docs/dyn/dataflow_v1b3.projects.html | 8 + .../dyn/dataflow_v1b3.projects.locations.html | 8 + ...projects.locations.connectionProfiles.html | 85 +- ...astream_v1.projects.locations.streams.html | 56 ++ ...dialogflow_v2beta1.projects.locations.html | 5 + ..._v2beta1.projects.locations.sipTrunks.html | 324 +++++++ ....projects.locations.agents.generators.html | 36 + ....projects.locations.agents.generators.html | 36 + ...llections.dataStores.siteSearchEngine.html | 5 + ....dataStores.siteSearchEngine.sitemaps.html | 171 ++++ ...locations.dataStores.siteSearchEngine.html | 5 + ....dataStores.siteSearchEngine.sitemaps.html | 202 ++++ ...llections.dataStores.siteSearchEngine.html | 5 + ....dataStores.siteSearchEngine.sitemaps.html | 171 ++++ ...locations.dataStores.siteSearchEngine.html | 5 + ....dataStores.siteSearchEngine.sitemaps.html | 202 ++++ docs/dyn/dlp_v2.infoTypes.html | 4 +- docs/dyn/dlp_v2.locations.infoTypes.html | 4 +- ...2.organizations.locations.connections.html | 28 +- ...anizations.locations.discoveryConfigs.html | 24 +- ...tions.locations.fileStoreDataProfiles.html | 16 +- ...nizations.locations.tableDataProfiles.html | 16 +- ...dlp_v2.projects.locations.connections.html | 28 +- ...2.projects.locations.discoveryConfigs.html | 24 +- ...jects.locations.fileStoreDataProfiles.html | 16 +- ....projects.locations.tableDataProfiles.html | 16 +- docs/dyn/drive_v3.files.accessproposals.html | 2 +- ...ntarc_v1.projects.locations.pipelines.html | 56 +- ...appdistribution_v1alpha.projects.apps.html | 9 + ..._v1alpha.projects.apps.releases.tests.html | 24 +- ...gkehub_v1.projects.locations.features.html | 20 + ...b_v1alpha.projects.locations.features.html | 12 + ...ub_v1beta.projects.locations.features.html | 20 + ...ojects.locations.memberships.features.html | 8 + ...ojects.locations.memberships.features.html | 8 + ...ojects.locations.memberships.features.html | 8 + ...rojects.locations.vmwareAdminClusters.html | 280 ++++++ ...accounts_v1beta.accounts.businessInfo.html | 12 +- ...api_accounts_v1beta.accounts.homepage.html | 10 +- .../merchantapi_accounts_v1beta.accounts.html | 16 +- ...antapi_accounts_v1beta.accounts.users.html | 12 +- ...antapi_accounts_v1beta.termsOfService.html | 4 +- ...ioncenter_v1alpha1.projects.locations.html | 2 + ...s.locations.global_.connectivityTests.html | 24 +- ...s.locations.global_.connectivityTests.html | 24 +- ...locations.global_.policyOrchestrators.html | 2 +- ...locations.global_.policyOrchestrators.html | 2 +- ...locations.global_.policyOrchestrators.html | 2 +- docs/dyn/places_v1.places.html | 6 +- ...chaenterprise_v1.projects.assessments.html | 4 +- .../recaptchaenterprise_v1.projects.keys.html | 2 +- ...tpu_v2alpha1.projects.locations.nodes.html | 76 ++ ...a1.projects.locations.queuedResources.html | 71 ++ docs/dyn/youtube_v3.playlistItems.html | 2 +- docs/dyn/youtube_v3.videos.html | 5 + .../documents/accesscontextmanager.v1.json | 10 +- .../documents/admin.directory_v1.json | 26 +- .../documents/aiplatform.v1.json | 390 +++++++- .../documents/aiplatform.v1beta1.json | 884 +++++++++++++++++- .../discovery_cache/documents/alloydb.v1.json | 4 +- .../documents/alloydb.v1alpha.json | 4 +- .../documents/alloydb.v1beta.json | 4 +- .../documents/analyticshub.v1.json | 6 +- .../documents/analyticshub.v1beta1.json | 94 +- .../documents/androidmanagement.v1.json | 4 +- .../documents/backupdr.v1.json | 4 +- .../documents/beyondcorp.v1.json | 74 +- .../discovery_cache/documents/chat.v1.json | 22 +- .../documents/clouddeploy.v1.json | 107 ++- .../documents/cloudfunctions.v2.json | 10 +- .../documents/cloudfunctions.v2alpha.json | 10 +- .../documents/cloudfunctions.v2beta.json | 10 +- .../documents/composer.v1.json | 18 +- .../documents/composer.v1beta1.json | 18 +- .../discovery_cache/documents/compute.v1.json | 64 +- .../documents/container.v1beta1.json | 70 +- .../documents/containeranalysis.v1alpha1.json | 8 +- .../documents/dataflow.v1b3.json | 37 +- .../documents/datastream.v1.json | 121 ++- .../documents/developerconnect.v1.json | 7 +- .../documents/dialogflow.v2beta1.json | 287 +++++- .../documents/dialogflow.v3.json | 33 +- .../documents/dialogflow.v3beta1.json | 33 +- .../documents/discoveryengine.v1.json | 164 +++- .../documents/discoveryengine.v1alpha.json | 309 +++++- .../documents/discoveryengine.v1beta.json | 309 +++++- .../documents/displayvideo.v2.json | 38 +- .../documents/displayvideo.v3.json | 44 +- .../discovery_cache/documents/dlp.v2.json | 20 +- .../discovery_cache/documents/drive.v3.json | 4 +- .../documents/eventarc.v1.json | 16 +- .../firebaseappdistribution.v1alpha.json | 16 +- .../discovery_cache/documents/gkehub.v1.json | 20 +- .../documents/gkehub.v1alpha.json | 6 +- .../documents/gkehub.v1beta.json | 20 +- .../discovery_cache/documents/gkehub.v2.json | 6 +- .../documents/gkehub.v2alpha.json | 6 +- .../documents/gkehub.v2beta.json | 6 +- .../documents/gkeonprem.v1.json | 45 +- .../merchantapi.accounts_v1beta.json | 14 +- .../merchantapi.conversions_v1beta.json | 9 +- .../documents/merchantapi.reports_v1beta.json | 16 +- .../documents/merchantapi.reviews_v1beta.json | 37 +- .../documents/migrationcenter.v1alpha1.json | 6 +- .../documents/networkmanagement.v1.json | 8 +- .../documents/networkmanagement.v1beta1.json | 8 +- .../documents/networkservices.v1beta1.json | 4 +- .../documents/osconfig.v2beta.json | 8 +- .../discovery_cache/documents/places.v1.json | 3 +- .../documents/recaptchaenterprise.v1.json | 6 +- .../documents/serviceusage.v1.json | 167 +++- .../documents/serviceusage.v1beta1.json | 167 +++- .../documents/tpu.v2alpha1.json | 140 ++- .../documents/translate.v3.json | 8 +- .../documents/translate.v3beta1.json | 11 +- .../discovery_cache/documents/youtube.v3.json | 8 +- 175 files changed, 8252 insertions(+), 733 deletions(-) create mode 100644 docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.html create mode 100644 docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.html create mode 100644 docs/dyn/dialogflow_v2beta1.projects.locations.sipTrunks.html create mode 100644 docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html create mode 100644 docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.sitemaps.html create mode 100644 docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html create mode 100644 docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.sitemaps.html diff --git a/docs/dyn/accesscontextmanager_v1.organizations.gcpUserAccessBindings.html b/docs/dyn/accesscontextmanager_v1.organizations.gcpUserAccessBindings.html index fe5699020db..9e863d6c824 100644 --- a/docs/dyn/accesscontextmanager_v1.organizations.gcpUserAccessBindings.html +++ b/docs/dyn/accesscontextmanager_v1.organizations.gcpUserAccessBindings.html @@ -110,7 +110,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. Next ID: 11 +{ # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. "accessLevels": [ # Optional. Access level that a user must have to be granted access. Only one access level is supported, not multiple. This repeated field must have exactly one element. Example: "accessPolicies/9522/accessLevels/device_trusted" "A String", ], @@ -182,6 +182,13 @@

Method Details

}, }, ], + "sessionSettings": { # Stores settings related to Google Cloud Session Length including session duration, the type of challenge (i.e. method) they should face when their session expires, and other related settings. # Optional. GCSL policy for the group key. Migrated from ReauthSettings + "maxInactivity": "A String", # Optional. How long a user is allowed to take between actions before a new access token must be issued. Presently only set for Cloud Apps. + "sessionLength": "A String", # Optional. The session length. Setting this field to zero is equal to disabling. Session. Also can set infinite session by flipping the enabled bit to false below. If use_oidc_max_age is true, for OIDC apps, the session length will be the minimum of this field and OIDC max_age param. + "sessionLengthEnabled": True or False, # Optional. Big red button to turn off GCSL. When false, all fields set above will be disregarded and the session length is basically infinite. + "sessionReauthMethod": "A String", # Optional. Session method when users GCP session is up. + "useOidcMaxAge": True or False, # Optional. Only useful for OIDC apps. When false, the OIDC max_age param, if passed in the authentication request will be ignored. When true, the re-auth period will be the minimum of the session_length field and the max_age OIDC param. + }, } x__xgafv: string, V1 error format. @@ -262,7 +269,7 @@

Method Details

Returns: An object of the form: - { # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. Next ID: 11 + { # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. "accessLevels": [ # Optional. Access level that a user must have to be granted access. Only one access level is supported, not multiple. This repeated field must have exactly one element. Example: "accessPolicies/9522/accessLevels/device_trusted" "A String", ], @@ -334,6 +341,13 @@

Method Details

}, }, ], + "sessionSettings": { # Stores settings related to Google Cloud Session Length including session duration, the type of challenge (i.e. method) they should face when their session expires, and other related settings. # Optional. GCSL policy for the group key. Migrated from ReauthSettings + "maxInactivity": "A String", # Optional. How long a user is allowed to take between actions before a new access token must be issued. Presently only set for Cloud Apps. + "sessionLength": "A String", # Optional. The session length. Setting this field to zero is equal to disabling. Session. Also can set infinite session by flipping the enabled bit to false below. If use_oidc_max_age is true, for OIDC apps, the session length will be the minimum of this field and OIDC max_age param. + "sessionLengthEnabled": True or False, # Optional. Big red button to turn off GCSL. When false, all fields set above will be disregarded and the session length is basically infinite. + "sessionReauthMethod": "A String", # Optional. Session method when users GCP session is up. + "useOidcMaxAge": True or False, # Optional. Only useful for OIDC apps. When false, the OIDC max_age param, if passed in the authentication request will be ignored. When true, the re-auth period will be the minimum of the session_length field and the max_age OIDC param. + }, } @@ -355,7 +369,7 @@

Method Details

{ # Response of ListGcpUserAccessBindings. "gcpUserAccessBindings": [ # GcpUserAccessBinding - { # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. Next ID: 11 + { # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. "accessLevels": [ # Optional. Access level that a user must have to be granted access. Only one access level is supported, not multiple. This repeated field must have exactly one element. Example: "accessPolicies/9522/accessLevels/device_trusted" "A String", ], @@ -427,6 +441,13 @@

Method Details

}, }, ], + "sessionSettings": { # Stores settings related to Google Cloud Session Length including session duration, the type of challenge (i.e. method) they should face when their session expires, and other related settings. # Optional. GCSL policy for the group key. Migrated from ReauthSettings + "maxInactivity": "A String", # Optional. How long a user is allowed to take between actions before a new access token must be issued. Presently only set for Cloud Apps. + "sessionLength": "A String", # Optional. The session length. Setting this field to zero is equal to disabling. Session. Also can set infinite session by flipping the enabled bit to false below. If use_oidc_max_age is true, for OIDC apps, the session length will be the minimum of this field and OIDC max_age param. + "sessionLengthEnabled": True or False, # Optional. Big red button to turn off GCSL. When false, all fields set above will be disregarded and the session length is basically infinite. + "sessionReauthMethod": "A String", # Optional. Session method when users GCP session is up. + "useOidcMaxAge": True or False, # Optional. Only useful for OIDC apps. When false, the OIDC max_age param, if passed in the authentication request will be ignored. When true, the re-auth period will be the minimum of the session_length field and the max_age OIDC param. + }, }, ], "nextPageToken": "A String", # Token to get the next page of items. If blank, there are no more items. @@ -456,7 +477,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. Next ID: 11 +{ # Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. "accessLevels": [ # Optional. Access level that a user must have to be granted access. Only one access level is supported, not multiple. This repeated field must have exactly one element. Example: "accessPolicies/9522/accessLevels/device_trusted" "A String", ], @@ -528,9 +549,16 @@

Method Details

}, }, ], + "sessionSettings": { # Stores settings related to Google Cloud Session Length including session duration, the type of challenge (i.e. method) they should face when their session expires, and other related settings. # Optional. GCSL policy for the group key. Migrated from ReauthSettings + "maxInactivity": "A String", # Optional. How long a user is allowed to take between actions before a new access token must be issued. Presently only set for Cloud Apps. + "sessionLength": "A String", # Optional. The session length. Setting this field to zero is equal to disabling. Session. Also can set infinite session by flipping the enabled bit to false below. If use_oidc_max_age is true, for OIDC apps, the session length will be the minimum of this field and OIDC max_age param. + "sessionLengthEnabled": True or False, # Optional. Big red button to turn off GCSL. When false, all fields set above will be disregarded and the session length is basically infinite. + "sessionReauthMethod": "A String", # Optional. Session method when users GCP session is up. + "useOidcMaxAge": True or False, # Optional. Only useful for OIDC apps. When false, the OIDC max_age param, if passed in the authentication request will be ignored. When true, the re-auth period will be the minimum of the session_length field and the max_age OIDC param. + }, } - append: boolean, Optional. This field controls whether or not certain repeated settings in the update request overwrite or append to existing settings on the binding. If true, then append. Otherwise overwrite. So far, only scoped_access_settings supports appending. Global access_levels, dry_run_access_levels, reauth_settings, and session_settings are not compatible with append functionality, and the request will return an error if append=true when these settings are in the update_mask. The request will also return an error if append=true when "scoped_access_settings" is not set in the update_mask. + append: boolean, Optional. This field controls whether or not certain repeated settings in the update request overwrite or append to existing settings on the binding. If true, then append. Otherwise overwrite. So far, only scoped_access_settings with reauth_settings supports appending. Global access_levels, access_levels in scoped_access_settings, dry_run_access_levels, reauth_settings, and session_settings are not compatible with append functionality, and the request will return an error if append=true when these settings are in the update_mask. The request will also return an error if append=true when "scoped_access_settings" is not set in the update_mask. updateMask: string, Required. Only the fields specified in this mask are updated. Because name and group_key cannot be changed, update_mask is required and may only contain the following fields: `access_levels`, `dry_run_access_levels`, `reauth_settings` `session_settings`, `scoped_access_settings`. update_mask { paths: "access_levels" } x__xgafv: string, V1 error format. Allowed values diff --git a/docs/dyn/admin_directory_v1.chromeosdevices.html b/docs/dyn/admin_directory_v1.chromeosdevices.html index 303526701a9..b9096291e09 100644 --- a/docs/dyn/admin_directory_v1.chromeosdevices.html +++ b/docs/dyn/admin_directory_v1.chromeosdevices.html @@ -211,6 +211,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes @@ -405,6 +409,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes @@ -614,6 +622,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes @@ -782,6 +794,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes @@ -949,6 +965,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes @@ -1117,6 +1137,10 @@

Method Details

], "deviceId": "A String", # The unique ID of the Chrome device. "deviceLicenseType": "A String", # Output only. Device license type. + "diskSpaceUsage": { # Represents a data capacity with some amount of current usage in bytes. # Output only. How much disk space the device has available and is currently using. + "capacityBytes": "A String", # Output only. The total capacity value, in bytes. + "usedBytes": "A String", # Output only. The current usage value, in bytes. + }, "diskVolumeReports": [ # Reports of disk space and other info about mounted/connected volumes. { "volumeInfo": [ # Disk volumes diff --git a/docs/dyn/aiplatform_v1.endpoints.html b/docs/dyn/aiplatform_v1.endpoints.html index dd47308d622..9872490666d 100644 --- a/docs/dyn/aiplatform_v1.endpoints.html +++ b/docs/dyn/aiplatform_v1.endpoints.html @@ -310,7 +310,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -584,7 +584,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1003,7 +1003,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1.projects.html b/docs/dyn/aiplatform_v1.projects.html index 47c211db216..d243f22e639 100644 --- a/docs/dyn/aiplatform_v1.projects.html +++ b/docs/dyn/aiplatform_v1.projects.html @@ -82,10 +82,79 @@

Instance Methods

close()

Close httplib2 connections.

+

+ getCacheConfig(name, x__xgafv=None)

+

Gets a GenAI cache config.

+

+ updateCacheConfig(name, body=None, x__xgafv=None)

+

Updates a cache config.

Method Details

close()
Close httplib2 connections.
+
+ getCacheConfig(name, x__xgafv=None) +
Gets a GenAI cache config.
+
+Args:
+  name: string, Required. Name of the cache config. Format: - `projects/{project}/cacheConfig`. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Config of GenAI caching features. This is a singleton resource.
+  "disableCache": True or False, # If set to true, disables GenAI caching. Otherwise caching is enabled.
+  "name": "A String", # Identifier. Name of the cache config. Format: - `projects/{project}/cacheConfig`.
+}
+
+ +
+ updateCacheConfig(name, body=None, x__xgafv=None) +
Updates a cache config.
+
+Args:
+  name: string, Identifier. Name of the cache config. Format: - `projects/{project}/cacheConfig`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Config of GenAI caching features. This is a singleton resource.
+  "disableCache": True or False, # If set to true, disables GenAI caching. Otherwise caching is enabled.
+  "name": "A String", # Identifier. Name of the cache config. Format: - `projects/{project}/cacheConfig`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ \ No newline at end of file diff --git a/docs/dyn/aiplatform_v1.projects.locations.endpoints.html b/docs/dyn/aiplatform_v1.projects.locations.endpoints.html index 7cc3be7b625..155c1426265 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.endpoints.html +++ b/docs/dyn/aiplatform_v1.projects.locations.endpoints.html @@ -111,6 +111,9 @@

Instance Methods

explain(endpoint, body=None, x__xgafv=None)

Perform an online explanation. If deployed_model_id is specified, the corresponding DeployModel must have explanation_spec populated. If deployed_model_id is not specified, all DeployedModels must have explanation_spec populated.

+

+ fetchPredictOperation(endpoint, body=None, x__xgafv=None)

+

Fetch an asynchronous online prediction operation.

generateContent(model, body=None, x__xgafv=None)

Generate content with multimodal inputs.

@@ -132,6 +135,9 @@

Instance Methods

predict(endpoint, body=None, x__xgafv=None)

Perform an online prediction.

+

+ predictLongRunning(endpoint, body=None, x__xgafv=None)

+

rawPredict(endpoint, body=None, x__xgafv=None)

Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.

@@ -368,7 +374,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1322,6 +1328,48 @@

Method Details

} +
+ fetchPredictOperation(endpoint, body=None, x__xgafv=None) +
Fetch an asynchronous online prediction operation.
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.FetchPredictOperation.
+  "operationName": "A String", # Required. The server-assigned name for the operation.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
generateContent(model, body=None, x__xgafv=None)
Generate content with multimodal inputs.
@@ -1478,7 +1526,7 @@ 

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -2853,6 +2901,51 @@

Method Details

}
+
+ predictLongRunning(endpoint, body=None, x__xgafv=None) +
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.PredictLongRunning.
+  "instances": [ # Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.
+    "",
+  ],
+  "parameters": "", # Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
rawPredict(endpoint, body=None, x__xgafv=None)
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.
@@ -3237,7 +3330,7 @@ 

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1.projects.locations.notebookExecutionJobs.html b/docs/dyn/aiplatform_v1.projects.locations.notebookExecutionJobs.html index 0b686962569..9d75b58f5b3 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.notebookExecutionJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.notebookExecutionJobs.html @@ -114,6 +114,30 @@

Method Details

{ # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -238,6 +262,30 @@

Method Details

{ # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -305,6 +353,30 @@

Method Details

"notebookExecutionJobs": [ # List of NotebookExecutionJobs in the requested page. { # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html index f2e0057a0eb..87ca98ab1e5 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html @@ -83,12 +83,18 @@

Instance Methods

countTokens(endpoint, body=None, x__xgafv=None)

Perform a token counting.

+

+ fetchPredictOperation(endpoint, body=None, x__xgafv=None)

+

Fetch an asynchronous online prediction operation.

generateContent(model, body=None, x__xgafv=None)

Generate content with multimodal inputs.

predict(endpoint, body=None, x__xgafv=None)

Perform an online prediction.

+

+ predictLongRunning(endpoint, body=None, x__xgafv=None)

+

rawPredict(endpoint, body=None, x__xgafv=None)

Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.

@@ -322,7 +328,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -440,6 +446,48 @@

Method Details

}
+
+ fetchPredictOperation(endpoint, body=None, x__xgafv=None) +
Fetch an asynchronous online prediction operation.
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.FetchPredictOperation.
+  "operationName": "A String", # Required. The server-assigned name for the operation.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
generateContent(model, body=None, x__xgafv=None)
Generate content with multimodal inputs.
@@ -596,7 +644,7 @@ 

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -895,6 +943,51 @@

Method Details

}
+
+ predictLongRunning(endpoint, body=None, x__xgafv=None) +
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.PredictLongRunning.
+  "instances": [ # Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.
+    "",
+  ],
+  "parameters": "", # Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
rawPredict(endpoint, body=None, x__xgafv=None)
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.
@@ -1279,7 +1372,7 @@ 

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1.projects.locations.schedules.html b/docs/dyn/aiplatform_v1.projects.locations.schedules.html index 88f89bf6a34..e7b8d3380b9 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.schedules.html +++ b/docs/dyn/aiplatform_v1.projects.locations.schedules.html @@ -127,6 +127,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -420,6 +444,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -755,6 +803,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1062,6 +1134,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1373,6 +1469,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1667,6 +1787,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html index dc32e7ba920..5321e98b736 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html @@ -140,7 +140,7 @@

Method Details

The object takes the form of: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. @@ -287,7 +287,7 @@

Method Details

An object of the form: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. @@ -441,7 +441,7 @@

Method Details

An object of the form: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. @@ -601,7 +601,7 @@

Method Details

"nextPageToken": "A String", # A token to retrieve the next page of results. Pass to ListTuningJobsRequest.page_token to obtain that page. "tuningJobs": [ # List of TuningJobs in the requested page. { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. @@ -776,7 +776,7 @@

Method Details

"tuningJob": "A String", # Support migration from tuning job list page, from gemini-1.0-pro-002 to 1.5 and above. }, "tuningJob": { # Represents a TuningJob that runs with Google owned models. # Optional. The TuningJob to be updated. Users can use this TuningJob field to overwrite tuning configs. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. diff --git a/docs/dyn/aiplatform_v1.publishers.models.html b/docs/dyn/aiplatform_v1.publishers.models.html index 7c62ad2de26..54bf5d5e5bd 100644 --- a/docs/dyn/aiplatform_v1.publishers.models.html +++ b/docs/dyn/aiplatform_v1.publishers.models.html @@ -313,7 +313,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -587,7 +587,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1399,7 +1399,7 @@

Method Details

"tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1beta1.endpoints.html b/docs/dyn/aiplatform_v1beta1.endpoints.html index 42588d36024..326b42a2b23 100644 --- a/docs/dyn/aiplatform_v1beta1.endpoints.html +++ b/docs/dyn/aiplatform_v1beta1.endpoints.html @@ -336,7 +336,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -629,7 +629,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1079,7 +1079,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html b/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html index 153f9db6c91..6a6229f4c26 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html @@ -211,7 +211,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -433,7 +433,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -680,7 +680,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -913,7 +913,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1154,7 +1154,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1377,7 +1377,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html index e04c6594575..7c387bd41d9 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html @@ -593,6 +593,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html b/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html index c01719e3051..85a6f68c4f8 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.endpoints.html @@ -111,6 +111,9 @@

Instance Methods

explain(endpoint, body=None, x__xgafv=None)

Perform an online explanation. If deployed_model_id is specified, the corresponding DeployModel must have explanation_spec populated. If deployed_model_id is not specified, all DeployedModels must have explanation_spec populated.

+

+ fetchPredictOperation(endpoint, body=None, x__xgafv=None)

+

Fetch an asynchronous online prediction operation.

generateContent(model, body=None, x__xgafv=None)

Generate content with multimodal inputs.

@@ -135,6 +138,9 @@

Instance Methods

predict(endpoint, body=None, x__xgafv=None)

Perform an online prediction.

+

+ predictLongRunning(endpoint, body=None, x__xgafv=None)

+

rawPredict(endpoint, body=None, x__xgafv=None)

Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.

@@ -406,7 +412,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -688,6 +694,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -962,6 +971,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -1499,6 +1511,48 @@

Method Details

}
+
+ fetchPredictOperation(endpoint, body=None, x__xgafv=None) +
Fetch an asynchronous online prediction operation.
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.FetchPredictOperation.
+  "operationName": "A String", # Required. The server-assigned name for the operation.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
generateContent(model, body=None, x__xgafv=None)
Generate content with multimodal inputs.
@@ -1674,7 +1728,7 @@ 

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -2118,6 +2172,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -2381,6 +2438,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -2608,6 +2668,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -2819,6 +2882,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -3035,6 +3101,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. @@ -3124,6 +3193,51 @@

Method Details

}
+
+ predictLongRunning(endpoint, body=None, x__xgafv=None) +
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.PredictLongRunning.
+  "instances": [ # Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.
+    "",
+  ],
+  "parameters": "", # Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
rawPredict(endpoint, body=None, x__xgafv=None)
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.
@@ -3585,7 +3699,7 @@ 

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -4132,6 +4246,9 @@

Method Details

}, }, }, + "fasterDeploymentConfig": { # Configuration for faster model deployment. # Configuration for faster model deployment. + "fastTryoutEnabled": True or False, # If true, enable fast tryout feature for this deployed model. + }, "id": "A String", # Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`. "model": "A String", # Required. The resource name of the Model that this is the deployment of. Note that the Model may be in a different location than the DeployedModel's Endpoint. The resource name may contain version id or version alias to specify the version. Example: `projects/{project}/locations/{location}/models/{model}@2` or `projects/{project}/locations/{location}/models/{model}@golden` if no version is specified, the default version will be deployed. "modelVersionId": "A String", # Output only. The version ID of the model that is deployed. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.extensions.html b/docs/dyn/aiplatform_v1beta1.projects.locations.extensions.html index f9435927d4d..e09a3f94990 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.extensions.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.extensions.html @@ -219,7 +219,7 @@

Method Details

"etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "extensionOperations": [ # Output only. Supported operations. { # Operation of an extension. - "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. + "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -379,7 +379,7 @@

Method Details

"etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "extensionOperations": [ # Output only. Supported operations. { # Operation of an extension. - "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. + "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -578,7 +578,7 @@

Method Details

"etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "extensionOperations": [ # Output only. Supported operations. { # Operation of an extension. - "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. + "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -755,7 +755,7 @@

Method Details

"etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "extensionOperations": [ # Output only. Supported operations. { # Operation of an extension. - "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. + "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -914,7 +914,7 @@

Method Details

"etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "extensionOperations": [ # Output only. Supported operations. { # Operation of an extension. - "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. + "functionDeclaration": { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. # Output only. Structured representation of a function declaration as defined by the OpenAPI Spec. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.html new file mode 100644 index 00000000000..07527a87929 --- /dev/null +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.html @@ -0,0 +1,346 @@ + + + +

Vertex AI API . projects . locations . featureGroups . featureMonitors . featureMonitorJobs

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, featureMonitorJobId=None, x__xgafv=None)

+

Creates a new feature monitor job.

+

+ get(name, x__xgafv=None)

+

Get a feature monitor job.

+

+ list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)

+

List feature monitor jobs.

+

+ list_next()

+

Retrieves the next page of results.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, featureMonitorJobId=None, x__xgafv=None) +
Creates a new feature monitor job.
+
+Args:
+  parent: string, Required. The resource name of FeatureMonitor to create FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Vertex AI Feature Monitor Job.
+  "createTime": "A String", # Output only. Timestamp when this FeatureMonitorJob was created. Creation of a FeatureMonitorJob means that the job is pending / waiting for sufficient resources but may not have started running yet.
+  "description": "A String", # Optional. Description of the FeatureMonitor.
+  "driftBaseFeatureMonitorJobId": "A String", # Output only. FeatureMonitorJob ID comparing to which the drift is calculated.
+  "driftBaseSnapshotTime": "A String", # Output only. Data snapshot time comparing to which the drift is calculated.
+  "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Output only. Feature selection config used when creating FeatureMonitorJob.
+    "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+      { # Feature configuration.
+        "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+        "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+      },
+    ],
+  },
+  "finalStatus": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Final status of the FeatureMonitorJob.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "jobSummary": { # Summary from the FeatureMonitorJob. # Output only. Summary from the FeatureMonitorJob.
+    "featureStatsAndAnomalies": [ # Output only. Features and their stats and anomalies
+      { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift.
+        "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence.
+        "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats.
+        "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold
+        "featureId": "A String", # Feature Id.
+        "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to.
+        "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly.
+        "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics.
+        "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats.
+      },
+    ],
+    "totalSlotMs": "A String", # Output only. BigQuery slot milliseconds consumed.
+  },
+  "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitorJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+    "a_key": "A String",
+  },
+  "name": "A String", # Identifier. Name of the FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`.
+  "triggerType": "A String", # Output only. Trigger type of the Feature Monitor Job.
+}
+
+  featureMonitorJobId: string, Optional. Output only. System-generated ID for feature monitor job.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Vertex AI Feature Monitor Job.
+  "createTime": "A String", # Output only. Timestamp when this FeatureMonitorJob was created. Creation of a FeatureMonitorJob means that the job is pending / waiting for sufficient resources but may not have started running yet.
+  "description": "A String", # Optional. Description of the FeatureMonitor.
+  "driftBaseFeatureMonitorJobId": "A String", # Output only. FeatureMonitorJob ID comparing to which the drift is calculated.
+  "driftBaseSnapshotTime": "A String", # Output only. Data snapshot time comparing to which the drift is calculated.
+  "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Output only. Feature selection config used when creating FeatureMonitorJob.
+    "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+      { # Feature configuration.
+        "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+        "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+      },
+    ],
+  },
+  "finalStatus": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Final status of the FeatureMonitorJob.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "jobSummary": { # Summary from the FeatureMonitorJob. # Output only. Summary from the FeatureMonitorJob.
+    "featureStatsAndAnomalies": [ # Output only. Features and their stats and anomalies
+      { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift.
+        "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence.
+        "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats.
+        "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold
+        "featureId": "A String", # Feature Id.
+        "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to.
+        "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly.
+        "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics.
+        "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats.
+      },
+    ],
+    "totalSlotMs": "A String", # Output only. BigQuery slot milliseconds consumed.
+  },
+  "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitorJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+    "a_key": "A String",
+  },
+  "name": "A String", # Identifier. Name of the FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`.
+  "triggerType": "A String", # Output only. Trigger type of the Feature Monitor Job.
+}
+
+ +
+ get(name, x__xgafv=None) +
Get a feature monitor job.
+
+Args:
+  name: string, Required. The name of the FeatureMonitorJob resource. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}` (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Vertex AI Feature Monitor Job.
+  "createTime": "A String", # Output only. Timestamp when this FeatureMonitorJob was created. Creation of a FeatureMonitorJob means that the job is pending / waiting for sufficient resources but may not have started running yet.
+  "description": "A String", # Optional. Description of the FeatureMonitor.
+  "driftBaseFeatureMonitorJobId": "A String", # Output only. FeatureMonitorJob ID comparing to which the drift is calculated.
+  "driftBaseSnapshotTime": "A String", # Output only. Data snapshot time comparing to which the drift is calculated.
+  "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Output only. Feature selection config used when creating FeatureMonitorJob.
+    "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+      { # Feature configuration.
+        "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+        "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+      },
+    ],
+  },
+  "finalStatus": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Final status of the FeatureMonitorJob.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "jobSummary": { # Summary from the FeatureMonitorJob. # Output only. Summary from the FeatureMonitorJob.
+    "featureStatsAndAnomalies": [ # Output only. Features and their stats and anomalies
+      { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift.
+        "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence.
+        "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats.
+        "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold
+        "featureId": "A String", # Feature Id.
+        "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to.
+        "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly.
+        "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics.
+        "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats.
+      },
+    ],
+    "totalSlotMs": "A String", # Output only. BigQuery slot milliseconds consumed.
+  },
+  "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitorJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+    "a_key": "A String",
+  },
+  "name": "A String", # Identifier. Name of the FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`.
+  "triggerType": "A String", # Output only. Trigger type of the Feature Monitor Job.
+}
+
+ +
+ list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None) +
List feature monitor jobs.
+
+Args:
+  parent: string, Required. The resource name of the FeatureMonitor to list FeatureMonitorJobs. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}` (required)
+  filter: string, Optional. Lists the FeatureMonitorJobs that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be Examples: * `create_time > "2020-01-01"` FeatureMonitorJobs created after 2020-01-01.
+  orderBy: string, Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields: * `create_time`
+  pageSize: integer, Optional. The maximum number of FeatureMonitorJobs to return. The service may return fewer than this value. If unspecified, at most 100 FeatureMonitorJobs will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
+  pageToken: string, Optional. A page token, received from a previous FeatureRegistryService.ListFeatureMonitorJobs call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to FeatureRegistryService.ListFeatureMonitorJobs must match the call that provided the page token.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response message for FeatureRegistryService.ListFeatureMonitorJobs.
+  "featureMonitorJobs": [ # The FeatureMonitorJobs matching the request.
+    { # Vertex AI Feature Monitor Job.
+      "createTime": "A String", # Output only. Timestamp when this FeatureMonitorJob was created. Creation of a FeatureMonitorJob means that the job is pending / waiting for sufficient resources but may not have started running yet.
+      "description": "A String", # Optional. Description of the FeatureMonitor.
+      "driftBaseFeatureMonitorJobId": "A String", # Output only. FeatureMonitorJob ID comparing to which the drift is calculated.
+      "driftBaseSnapshotTime": "A String", # Output only. Data snapshot time comparing to which the drift is calculated.
+      "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Output only. Feature selection config used when creating FeatureMonitorJob.
+        "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+          { # Feature configuration.
+            "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+            "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+          },
+        ],
+      },
+      "finalStatus": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Final status of the FeatureMonitorJob.
+        "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+        "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+          {
+            "a_key": "", # Properties of the object. Contains field @type with type URL.
+          },
+        ],
+        "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+      },
+      "jobSummary": { # Summary from the FeatureMonitorJob. # Output only. Summary from the FeatureMonitorJob.
+        "featureStatsAndAnomalies": [ # Output only. Features and their stats and anomalies
+          { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift.
+            "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence.
+            "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats.
+            "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold
+            "featureId": "A String", # Feature Id.
+            "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to.
+            "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly.
+            "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics.
+            "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats.
+          },
+        ],
+        "totalSlotMs": "A String", # Output only. BigQuery slot milliseconds consumed.
+      },
+      "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitorJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+        "a_key": "A String",
+      },
+      "name": "A String", # Identifier. Name of the FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`.
+      "triggerType": "A String", # Output only. Trigger type of the Feature Monitor Job.
+    },
+  ],
+  "nextPageToken": "A String", # A token, which can be sent as ListFeatureMonitorJobsRequest.page_token to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+ +
+ list_next() +
Retrieves the next page of results.
+
+        Args:
+          previous_request: The request for the previous page. (required)
+          previous_response: The response from the request for the previous page. (required)
+
+        Returns:
+          A request object that you can call 'execute()' on to request the next
+          page. Returns None if there are no more items in the collection.
+        
+
+ + \ No newline at end of file diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.html new file mode 100644 index 00000000000..ca5df257eff --- /dev/null +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.featureMonitors.html @@ -0,0 +1,299 @@ + + + +

Vertex AI API . projects . locations . featureGroups . featureMonitors

+

Instance Methods

+

+ featureMonitorJobs() +

+

Returns the featureMonitorJobs Resource.

+ +

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, featureMonitorId=None, x__xgafv=None)

+

Creates a new FeatureMonitor in a given project, location and FeatureGroup.

+

+ delete(name, x__xgafv=None)

+

Deletes a single FeatureMonitor.

+

+ get(name, x__xgafv=None)

+

Gets details of a single FeatureMonitor.

+

+ list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None)

+

Lists FeatureGroups in a given project and location.

+

+ list_next()

+

Retrieves the next page of results.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, featureMonitorId=None, x__xgafv=None) +
Creates a new FeatureMonitor in a given project, location and FeatureGroup.
+
+Args:
+  parent: string, Required. The resource name of FeatureGroup to create FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featuregroup}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Vertex AI Feature Monitor.
+  "createTime": "A String", # Output only. Timestamp when this FeatureMonitor was created.
+  "description": "A String", # Optional. Description of the FeatureMonitor.
+  "etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
+  "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Required. Feature selection config for the FeatureMonitor.
+    "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+      { # Feature configuration.
+        "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+        "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+      },
+    ],
+  },
+  "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitor. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+    "a_key": "A String",
+  },
+  "name": "A String", # Identifier. Name of the FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}/featureMonitors/{featureMonitor}`
+  "scheduleConfig": { # Schedule configuration for the FeatureMonitor. # Required. Schedule config for the FeatureMonitor.
+    "cron": "A String", # Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
+  },
+  "updateTime": "A String", # Output only. Timestamp when this FeatureMonitor was last updated.
+}
+
+  featureMonitorId: string, Required. The ID to use for this FeatureMonitor, which will become the final component of the FeatureGroup's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within the FeatureGroup.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a single FeatureMonitor.
+
+Args:
+  name: string, Required. The name of the FeatureMonitor to be deleted. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}` (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ get(name, x__xgafv=None) +
Gets details of a single FeatureMonitor.
+
+Args:
+  name: string, Required. The name of the FeatureMonitor resource. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Vertex AI Feature Monitor.
+  "createTime": "A String", # Output only. Timestamp when this FeatureMonitor was created.
+  "description": "A String", # Optional. Description of the FeatureMonitor.
+  "etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
+  "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Required. Feature selection config for the FeatureMonitor.
+    "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+      { # Feature configuration.
+        "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+        "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+      },
+    ],
+  },
+  "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitor. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+    "a_key": "A String",
+  },
+  "name": "A String", # Identifier. Name of the FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}/featureMonitors/{featureMonitor}`
+  "scheduleConfig": { # Schedule configuration for the FeatureMonitor. # Required. Schedule config for the FeatureMonitor.
+    "cron": "A String", # Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
+  },
+  "updateTime": "A String", # Output only. Timestamp when this FeatureMonitor was last updated.
+}
+
+ +
+ list(parent, filter=None, orderBy=None, pageSize=None, pageToken=None, x__xgafv=None) +
Lists FeatureGroups in a given project and location.
+
+Args:
+  parent: string, Required. The resource name of the FeatureGroup to list FeatureMonitors. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}` (required)
+  filter: string, Optional. Lists the FeatureMonitors that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality and key presence. Examples: * `create_time > "2020-01-01" OR update_time > "2020-01-01"` FeatureMonitors created or updated after 2020-01-01. * `labels.env = "prod"` FeatureGroups with label "env" set to "prod".
+  orderBy: string, Optional. A comma-separated list of fields to order by, sorted in ascending order. Use "desc" after a field name for descending. Supported Fields: * `create_time` * `update_time`
+  pageSize: integer, Optional. The maximum number of FeatureGroups to return. The service may return fewer than this value. If unspecified, at most 100 FeatureMonitors will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.
+  pageToken: string, Optional. A page token, received from a previous FeatureRegistryService.ListFeatureMonitors call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to FeatureRegistryService.ListFeatureMonitors must match the call that provided the page token.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response message for FeatureRegistryService.ListFeatureMonitors.
+  "featureMonitors": [ # The FeatureMonitors matching the request.
+    { # Vertex AI Feature Monitor.
+      "createTime": "A String", # Output only. Timestamp when this FeatureMonitor was created.
+      "description": "A String", # Optional. Description of the FeatureMonitor.
+      "etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens.
+      "featureSelectionConfig": { # Feature selection configuration for the FeatureMonitor. # Required. Feature selection config for the FeatureMonitor.
+        "featureConfigs": [ # Optional. A list of features to be monitored and each feature's drift threshold.
+          { # Feature configuration.
+            "driftThreshold": 3.14, # Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.
+            "featureId": "A String", # Required. The ID of the feature resource. Final component of the Feature's resource name.
+          },
+        ],
+      },
+      "labels": { # Optional. The labels with user-defined metadata to organize your FeatureMonitor. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable.
+        "a_key": "A String",
+      },
+      "name": "A String", # Identifier. Name of the FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}/featureMonitors/{featureMonitor}`
+      "scheduleConfig": { # Schedule configuration for the FeatureMonitor. # Required. Schedule config for the FeatureMonitor.
+        "cron": "A String", # Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: "CRON_TZ=${IANA_TIME_ZONE}" or "TZ=${IANA_TIME_ZONE}". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *".
+      },
+      "updateTime": "A String", # Output only. Timestamp when this FeatureMonitor was last updated.
+    },
+  ],
+  "nextPageToken": "A String", # A token, which can be sent as ListFeatureMonitorsRequest.page_token to retrieve the next page. If this field is omitted, there are no subsequent pages.
+}
+
+ +
+ list_next() +
Retrieves the next page of results.
+
+        Args:
+          previous_request: The request for the previous page. (required)
+          previous_response: The response from the request for the previous page. (required)
+
+        Returns:
+          A request object that you can call 'execute()' on to request the next
+          page. Returns None if there are no more items in the collection.
+        
+
+ + \ No newline at end of file diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.features.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.features.html index cc4e79f3969..231c110ad2e 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.features.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.features.html @@ -92,7 +92,7 @@

Instance Methods

delete(name, x__xgafv=None)

Deletes a single Feature.

- get(name, x__xgafv=None)

+ get(name, featureStatsAndAnomalySpec_latestStatsCount=None, featureStatsAndAnomalySpec_statsTimeRange_endTime=None, featureStatsAndAnomalySpec_statsTimeRange_startTime=None, x__xgafv=None)

Gets details of a single Feature.

list(parent, filter=None, latestStatsCount=None, orderBy=None, pageSize=None, pageToken=None, readMask=None, x__xgafv=None)

@@ -121,6 +121,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -227,6 +239,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -346,11 +370,14 @@

Method Details

- get(name, x__xgafv=None) + get(name, featureStatsAndAnomalySpec_latestStatsCount=None, featureStatsAndAnomalySpec_statsTimeRange_endTime=None, featureStatsAndAnomalySpec_statsTimeRange_startTime=None, x__xgafv=None)
Gets details of a single Feature.
 
 Args:
   name: string, Required. The name of the Feature resource. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}` (required)
+  featureStatsAndAnomalySpec_latestStatsCount: integer, Optional. If set, returns the most recent count of stats. Valid value is [0, 100]. If stats_time_range is set, return most recent count of stats within the stats_time_range.
+  featureStatsAndAnomalySpec_statsTimeRange_endTime: string, Optional. Exclusive end of the interval. If specified, a Timestamp matching this interval will have to be before the end.
+  featureStatsAndAnomalySpec_statsTimeRange_startTime: string, Optional. Inclusive start of the interval. If specified, a Timestamp matching this interval will have to be the same or after the start.
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -364,6 +391,18 @@ 

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -445,6 +484,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -530,6 +581,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.html index a5fcb36647a..8567f3d238c 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featureGroups.html @@ -74,6 +74,11 @@

Vertex AI API . projects . locations . featureGroups

Instance Methods

+

+ featureMonitors() +

+

Returns the featureMonitors Resource.

+

features()

diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.features.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.features.html index 3a473505188..7f66acdbb0b 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.features.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.entityTypes.features.html @@ -92,7 +92,7 @@

Instance Methods

delete(name, x__xgafv=None)

Deletes a single Feature.

- get(name, x__xgafv=None)

+ get(name, featureStatsAndAnomalySpec_latestStatsCount=None, featureStatsAndAnomalySpec_statsTimeRange_endTime=None, featureStatsAndAnomalySpec_statsTimeRange_startTime=None, x__xgafv=None)

Gets details of a single Feature.

list(parent, filter=None, latestStatsCount=None, orderBy=None, pageSize=None, pageToken=None, readMask=None, x__xgafv=None)

@@ -121,6 +121,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -227,6 +239,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -346,11 +370,14 @@

Method Details

- get(name, x__xgafv=None) + get(name, featureStatsAndAnomalySpec_latestStatsCount=None, featureStatsAndAnomalySpec_statsTimeRange_endTime=None, featureStatsAndAnomalySpec_statsTimeRange_startTime=None, x__xgafv=None)
Gets details of a single Feature.
 
 Args:
   name: string, Required. The name of the Feature resource. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}` (required)
+  featureStatsAndAnomalySpec_latestStatsCount: integer, Optional. If set, returns the most recent count of stats. Valid value is [0, 100]. If stats_time_range is set, return most recent count of stats within the stats_time_range.
+  featureStatsAndAnomalySpec_statsTimeRange_endTime: string, Optional. Exclusive end of the interval. If specified, a Timestamp matching this interval will have to be before the end.
+  featureStatsAndAnomalySpec_statsTimeRange_startTime: string, Optional. Inclusive start of the interval. If specified, a Timestamp matching this interval will have to be the same or after the start.
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -364,6 +391,18 @@ 

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -445,6 +484,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -530,6 +581,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, @@ -597,6 +660,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.html b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.html index 20eb9c5b437..21af1b8e3e4 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.featurestores.html @@ -554,6 +554,18 @@

Method Details

"description": "A String", # Description of the Feature. "disableMonitoring": True or False, # Optional. Only applicable for Vertex AI Feature Store (Legacy). If not set, use the monitoring_config defined for the EntityType this Feature belongs to. Only Features with type (Feature.ValueType) BOOL, STRING, DOUBLE or INT64 can enable monitoring. If set to true, all types of data monitoring are disabled despite the config on EntityType. "etag": "A String", # Used to perform a consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "featureStatsAndAnomaly": [ # Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies. + { # Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift. + "distributionDeviation": 3.14, # Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen–Shannon divergence. + "driftDetected": True or False, # If set to true, indicates current stats is detected as and comparing with baseline stats. + "driftDetectionThreshold": 3.14, # This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold + "featureId": "A String", # Feature Id. + "featureMonitorId": "A String", # The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to. + "featureMonitorJobId": "A String", # The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly. + "featureStats": "", # Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics. + "statsTime": "A String", # The timestamp we take snapshot for feature values to generate stats. + }, + ], "labels": { # Optional. The labels with user-defined metadata to organize your Features. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one Feature (System labels are excluded)." System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. "a_key": "A String", }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookExecutionJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookExecutionJobs.html index 8332d24867b..63eee402f2a 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookExecutionJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookExecutionJobs.html @@ -120,6 +120,30 @@

Method Details

{ # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -273,6 +297,30 @@

Method Details

{ # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -340,6 +388,30 @@

Method Details

"notebookExecutionJobs": [ # List of NotebookExecutionJobs in the requested page. { # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html index e5c1b1a7f19..dfc602f26e9 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html @@ -83,6 +83,9 @@

Instance Methods

countTokens(endpoint, body=None, x__xgafv=None)

Perform a token counting.

+

+ fetchPredictOperation(endpoint, body=None, x__xgafv=None)

+

Fetch an asynchronous online prediction operation.

generateContent(model, body=None, x__xgafv=None)

Generate content with multimodal inputs.

@@ -92,6 +95,9 @@

Instance Methods

predict(endpoint, body=None, x__xgafv=None)

Perform an online prediction.

+

+ predictLongRunning(endpoint, body=None, x__xgafv=None)

+

rawPredict(endpoint, body=None, x__xgafv=None)

Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.

@@ -351,7 +357,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -469,6 +475,48 @@

Method Details

}
+
+ fetchPredictOperation(endpoint, body=None, x__xgafv=None) +
Fetch an asynchronous online prediction operation.
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.FetchPredictOperation.
+  "operationName": "A String", # Required. The server-assigned name for the operation.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
generateContent(model, body=None, x__xgafv=None)
Generate content with multimodal inputs.
@@ -644,7 +692,7 @@ 

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -990,6 +1038,51 @@

Method Details

}
+
+ predictLongRunning(endpoint, body=None, x__xgafv=None) +
+
+Args:
+  endpoint: string, Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}` (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for PredictionService.PredictLongRunning.
+  "instances": [ # Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.
+    "",
+  ],
+  "parameters": "", # Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
rawPredict(endpoint, body=None, x__xgafv=None)
Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.
@@ -1393,7 +1486,7 @@ 

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html index c4ffbfe5eab..522a500ef9b 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html @@ -371,6 +371,7 @@

Method Details

The object takes the form of: { # Request message for ReasoningEngineExecutionService.Query. + "classMethod": "A String", # Optional. Class method to be used for the query. It is optional and defaults to "query" if unspecified. "input": { # Optional. Input content provided by users in JSON object format. Examples include text query, function calling parameters, media bytes, etc. "a_key": "", # Properties of the object. }, diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html index f13e6250643..3e300ebab22 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html @@ -443,6 +443,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1111,6 +1135,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1821,6 +1869,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -2503,6 +2575,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -3189,6 +3285,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -3858,6 +3978,30 @@

Method Details

"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. + "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "reservationAffinity": { # A ReservationAffinity can be used to configure a Vertex AI resource (e.g., a DeployedModel) to draw its Compute Engine resources from a Shared Reservation, or exclusively from on-demand capacity. # Optional. Immutable. Configuration controlling how this resource pool consumes reservation. + "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a SPECIFIC_RESERVATION by name, use `compute.googleapis.com/reservation-name` as the key and specify the name of your reservation as its value. + "reservationAffinityType": "A String", # Required. Specifies the reservation affinity type. + "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full resource name of the reservation. + "A String", + ], + }, + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "networkSpec": { # Network spec. # The network configuration to use for the execution job. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html index ac215f13aae..a536182b35a 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html @@ -140,7 +140,7 @@

Method Details

The object takes the form of: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. @@ -414,7 +414,7 @@

Method Details

An object of the form: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. @@ -695,7 +695,7 @@

Method Details

An object of the form: { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. @@ -982,7 +982,7 @@

Method Details

"nextPageToken": "A String", # A token to retrieve the next page of results. Pass to ListTuningJobsRequest.page_token to obtain that page. "tuningJobs": [ # List of TuningJobs in the requested page. { # Represents a TuningJob that runs with Google owned models. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. @@ -1284,7 +1284,7 @@

Method Details

"tuningJob": "A String", # Support migration from tuning job list page, from gemini-1.0-pro-002 to 1.5 and above. }, "tuningJob": { # Represents a TuningJob that runs with Google owned models. # Optional. The TuningJob to be updated. Users can use this TuningJob field to overwrite tuning configs. - "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". + "baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". . "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. diff --git a/docs/dyn/aiplatform_v1beta1.publishers.models.html b/docs/dyn/aiplatform_v1beta1.publishers.models.html index 5fbb7e98826..b41b8db2745 100644 --- a/docs/dyn/aiplatform_v1beta1.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.publishers.models.html @@ -345,7 +345,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -638,7 +638,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 @@ -1913,7 +1913,7 @@

Method Details

"codeExecution": { # Tool that executes code generated by the model, and automatically returns the result to the model. See also [ExecutableCode]and [CodeExecutionResult] which are input and output to this tool. # Optional. CodeExecution tool type. Enables the model to execute code as part of generation. This field is only used by the Gemini Developer API services. }, "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 128 function declarations can be provided. - { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema-object). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 diff --git a/docs/dyn/alloydb_v1.projects.locations.clusters.instances.html b/docs/dyn/alloydb_v1.projects.locations.clusters.instances.html index a3538088d33..96e8dcc37b4 100644 --- a/docs/dyn/alloydb_v1.projects.locations.clusters.instances.html +++ b/docs/dyn/alloydb_v1.projects.locations.clusters.instances.html @@ -138,7 +138,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -258,7 +258,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -469,7 +469,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -636,7 +636,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -745,7 +745,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp diff --git a/docs/dyn/alloydb_v1alpha.projects.locations.clusters.instances.html b/docs/dyn/alloydb_v1alpha.projects.locations.clusters.instances.html index a2ee69c5538..3512ff46b56 100644 --- a/docs/dyn/alloydb_v1alpha.projects.locations.clusters.instances.html +++ b/docs/dyn/alloydb_v1alpha.projects.locations.clusters.instances.html @@ -138,7 +138,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -281,7 +281,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -515,7 +515,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -709,7 +709,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -841,7 +841,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp diff --git a/docs/dyn/alloydb_v1beta.projects.locations.clusters.instances.html b/docs/dyn/alloydb_v1beta.projects.locations.clusters.instances.html index 1b13055f332..8dc3cc2384f 100644 --- a/docs/dyn/alloydb_v1beta.projects.locations.clusters.instances.html +++ b/docs/dyn/alloydb_v1beta.projects.locations.clusters.instances.html @@ -138,7 +138,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -275,7 +275,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -503,7 +503,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -691,7 +691,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp @@ -817,7 +817,7 @@

Method Details

}, }, "createTime": "A String", # Output only. Create time stamp - "databaseFlags": { # Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. + "databaseFlags": { # Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of "key": "value" pairs. "key": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. "value": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value. "a_key": "A String", }, "deleteTime": "A String", # Output only. Delete time stamp diff --git a/docs/dyn/analyticshub_v1.projects.locations.dataExchanges.html b/docs/dyn/analyticshub_v1.projects.locations.dataExchanges.html index 24b8bd34281..6bf17ea1b95 100644 --- a/docs/dyn/analyticshub_v1.projects.locations.dataExchanges.html +++ b/docs/dyn/analyticshub_v1.projects.locations.dataExchanges.html @@ -561,6 +561,18 @@

Method Details

{ # Message for subscribing to a Data Exchange. "destination": "A String", # Required. The parent resource path of the Subscription. e.g. `projects/subscriberproject/locations/US` + "destinationDataset": { # Defines the destination bigquery dataset. # Optional. BigQuery destination dataset to create for the subscriber. + "datasetReference": { # Contains the reference that identifies a destination bigquery dataset. # Required. A reference that identifies the destination dataset. + "datasetId": "A String", # Required. A unique ID for this dataset, without the project name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters. + "projectId": "A String", # Required. The ID of the project containing this dataset. + }, + "description": "A String", # Optional. A user-friendly description of the dataset. + "friendlyName": "A String", # Optional. A descriptive name for the dataset. + "labels": { # Optional. The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See https://cloud.google.com/resource-manager/docs/creating-managing-labels for more information. + "a_key": "A String", + }, + "location": "A String", # Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations. + }, "subscriberContact": "A String", # Email of the subscriber. "subscription": "A String", # Required. Name of the subscription to create. e.g. `subscription1` } diff --git a/docs/dyn/androidmanagement_v1.enterprises.policies.html b/docs/dyn/androidmanagement_v1.enterprises.policies.html index 29eb73d692e..e4d10c6e19e 100644 --- a/docs/dyn/androidmanagement_v1.enterprises.policies.html +++ b/docs/dyn/androidmanagement_v1.enterprises.policies.html @@ -169,7 +169,7 @@

Method Details

"A String", ], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. - "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 13 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 13 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. + "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hash of the signing certificate of the extension app. Only hexadecimal string representations of 64 characters are valid.If not specified, the signature for the corresponding package name is obtained from the Play Store instead.If this list is empty, the signature of the extension app on the device must match the signature obtained from the Play Store for the app to be able to communicate with Android Device Policy.If this list is not empty, the signature of the extension app on the device must match one of the entries in this list for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", @@ -609,7 +609,7 @@

Method Details

"A String", ], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. - "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 13 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 13 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. + "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hash of the signing certificate of the extension app. Only hexadecimal string representations of 64 characters are valid.If not specified, the signature for the corresponding package name is obtained from the Play Store instead.If this list is empty, the signature of the extension app on the device must match the signature obtained from the Play Store for the app to be able to communicate with Android Device Policy.If this list is not empty, the signature of the extension app on the device must match one of the entries in this list for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", @@ -1055,7 +1055,7 @@

Method Details

"A String", ], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. - "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 13 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 13 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. + "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hash of the signing certificate of the extension app. Only hexadecimal string representations of 64 characters are valid.If not specified, the signature for the corresponding package name is obtained from the Play Store instead.If this list is empty, the signature of the extension app on the device must match the signature obtained from the Play Store for the app to be able to communicate with Android Device Policy.If this list is not empty, the signature of the extension app on the device must match one of the entries in this list for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", @@ -1484,7 +1484,7 @@

Method Details

"A String", ], "disabled": True or False, # Whether the app is disabled. When disabled, the app data is still preserved. - "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 13 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 13 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. + "extensionConfig": { # Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above. # Configuration to enable this app as an extension app, with the capability of interacting with Android Device Policy offline.This field can be set for at most one app. "notificationReceiver": "A String", # Fully qualified class name of the receiver service class for Android Device Policy to notify the extension app of any local command status updates. "signingKeyFingerprintsSha256": [ # Hex-encoded SHA-256 hash of the signing certificate of the extension app. Only hexadecimal string representations of 64 characters are valid.If not specified, the signature for the corresponding package name is obtained from the Play Store instead.If this list is empty, the signature of the extension app on the device must match the signature obtained from the Play Store for the app to be able to communicate with Android Device Policy.If this list is not empty, the signature of the extension app on the device must match one of the entries in this list for the app to be able to communicate with Android Device Policy.In production use cases, it is recommended to leave this empty. "A String", diff --git a/docs/dyn/backupdr_v1.projects.locations.managementServers.html b/docs/dyn/backupdr_v1.projects.locations.managementServers.html index 55d95eb6d4d..c8d1df06113 100644 --- a/docs/dyn/backupdr_v1.projects.locations.managementServers.html +++ b/docs/dyn/backupdr_v1.projects.locations.managementServers.html @@ -131,7 +131,7 @@

Method Details

"webUi": "A String", # Output only. The ManagementServer AGM/RD WebUI URL. }, "name": "A String", # Output only. Identifier. The resource name. - "networks": [ # Required. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. + "networks": [ # Optional. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. This field is optional if MS is created without PSA { # Network configuration for ManagementServer instance. "network": "A String", # Optional. The resource name of the Google Compute Engine VPC network to which the ManagementServer instance is connected. "peeringMode": "A String", # Optional. The network connect mode of the ManagementServer instance. For this version, only PRIVATE_SERVICE_ACCESS is supported. @@ -249,7 +249,7 @@

Method Details

"webUi": "A String", # Output only. The ManagementServer AGM/RD WebUI URL. }, "name": "A String", # Output only. Identifier. The resource name. - "networks": [ # Required. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. + "networks": [ # Optional. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. This field is optional if MS is created without PSA { # Network configuration for ManagementServer instance. "network": "A String", # Optional. The resource name of the Google Compute Engine VPC network to which the ManagementServer instance is connected. "peeringMode": "A String", # Optional. The network connect mode of the ManagementServer instance. For this version, only PRIVATE_SERVICE_ACCESS is supported. @@ -355,7 +355,7 @@

Method Details

"webUi": "A String", # Output only. The ManagementServer AGM/RD WebUI URL. }, "name": "A String", # Output only. Identifier. The resource name. - "networks": [ # Required. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. + "networks": [ # Optional. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. This field is optional if MS is created without PSA { # Network configuration for ManagementServer instance. "network": "A String", # Optional. The resource name of the Google Compute Engine VPC network to which the ManagementServer instance is connected. "peeringMode": "A String", # Optional. The network connect mode of the ManagementServer instance. For this version, only PRIVATE_SERVICE_ACCESS is supported. diff --git a/docs/dyn/beyondcorp_v1.html b/docs/dyn/beyondcorp_v1.html index fe7bf91dec4..75b3d67705b 100644 --- a/docs/dyn/beyondcorp_v1.html +++ b/docs/dyn/beyondcorp_v1.html @@ -84,11 +84,6 @@

Instance Methods

Returns the projects Resource.

-

- v() -

-

Returns the v Resource.

-

close()

Close httplib2 connections.

diff --git a/docs/dyn/beyondcorp_v1.projects.locations.securityGateways.html b/docs/dyn/beyondcorp_v1.projects.locations.securityGateways.html index b06ad15494c..067b2f5588b 100644 --- a/docs/dyn/beyondcorp_v1.projects.locations.securityGateways.html +++ b/docs/dyn/beyondcorp_v1.projects.locations.securityGateways.html @@ -103,6 +103,9 @@

Instance Methods

patch(name, body=None, requestId=None, updateMask=None, x__xgafv=None)

Updates the parameters of a single SecurityGateway.

+

+ setIamPolicy(resource, body=None, x__xgafv=None)

+

Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.

setPeering(securityGateway, body=None, x__xgafv=None)

This is a custom method to allow customers to create a peering connections between Google network and customer networks. This is enabled only for the allowlisted customers.

@@ -416,6 +419,91 @@

Method Details

}
+
+ setIamPolicy(resource, body=None, x__xgafv=None) +
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
+
+Args:
+  resource: string, REQUIRED: The resource for which the policy is being specified. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request message for `SetIamPolicy` method.
+  "policy": { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** ``` { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } ``` **YAML example:** ``` bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 ``` For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/). # REQUIRED: The complete policy to be applied to the `resource`. The size of the policy is limited to a few 10s of KB. An empty policy is a valid policy but certain Google Cloud services (such as Projects) might reject them.
+    "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+      { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts `jose@example.com` from DATA_READ logging, and `aliya@example.com` from DATA_WRITE logging.
+        "auditLogConfigs": [ # The configuration for logging of each type of permission.
+          { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+            "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+              "A String",
+            ],
+            "logType": "A String", # The log type that this config enables.
+          },
+        ],
+        "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+      },
+    ],
+    "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`.
+      { # Associates `members`, or principals, with a `role`.
+        "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+          "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+          "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+          "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+          "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+        },
+        "members": [ # Specifies the principals requesting access for a Google Cloud resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. Does not include identities that come from external identity providers (IdPs) through identity federation. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a Google service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `serviceAccount:{projectid}.svc.id.goog[{namespace}/{kubernetes-sa}]`: An identifier for a [Kubernetes service account](https://cloud.google.com/kubernetes-engine/docs/how-to/kubernetes-service-accounts). For example, `my-project.svc.id.goog[my-namespace/my-kubernetes-sa]`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. * `principal://iam.googleapis.com/locations/global/workforcePools/{pool_id}/subject/{subject_attribute_value}`: A single identity in a workforce identity pool. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/group/{group_id}`: All workforce identities in a group. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/attribute.{attribute_name}/{attribute_value}`: All workforce identities with a specific attribute value. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/*`: All identities in a workforce identity pool. * `principal://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/subject/{subject_attribute_value}`: A single identity in a workload identity pool. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/group/{group_id}`: A workload identity pool group. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/attribute.{attribute_name}/{attribute_value}`: All identities in a workload identity pool with a certain attribute. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/*`: All identities in a workload identity pool. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `deleted:principal://iam.googleapis.com/locations/global/workforcePools/{pool_id}/subject/{subject_attribute_value}`: Deleted single identity in a workforce identity pool. For example, `deleted:principal://iam.googleapis.com/locations/global/workforcePools/my-pool-id/subject/my-subject-attribute-value`.
+          "A String",
+        ],
+        "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. For an overview of the IAM roles and permissions, see the [IAM documentation](https://cloud.google.com/iam/docs/roles-overview). For a list of the available pre-defined roles, see [here](https://cloud.google.com/iam/docs/understanding-roles).
+      },
+    ],
+    "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+    "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+  },
+  "updateMask": "A String", # OPTIONAL: A FieldMask specifying which fields of the policy to modify. Only the fields in the mask will be modified. If no mask is provided, the following default mask is used: `paths: "bindings, etag"`
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # An Identity and Access Management (IAM) policy, which specifies access controls for Google Cloud resources. A `Policy` is a collection of `bindings`. A `binding` binds one or more `members`, or principals, to a single `role`. Principals can be user accounts, service accounts, Google groups, and domains (such as G Suite). A `role` is a named list of permissions; each `role` can be an IAM predefined role or a user-created custom role. For some types of Google Cloud resources, a `binding` can also specify a `condition`, which is a logical expression that allows access to a resource only if the expression evaluates to `true`. A condition can add constraints based on attributes of the request, the resource, or both. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies). **JSON example:** ``` { "bindings": [ { "role": "roles/resourcemanager.organizationAdmin", "members": [ "user:mike@example.com", "group:admins@example.com", "domain:google.com", "serviceAccount:my-project-id@appspot.gserviceaccount.com" ] }, { "role": "roles/resourcemanager.organizationViewer", "members": [ "user:eve@example.com" ], "condition": { "title": "expirable access", "description": "Does not grant access after Sep 2020", "expression": "request.time < timestamp('2020-10-01T00:00:00.000Z')", } } ], "etag": "BwWWja0YfJA=", "version": 3 } ``` **YAML example:** ``` bindings: - members: - user:mike@example.com - group:admins@example.com - domain:google.com - serviceAccount:my-project-id@appspot.gserviceaccount.com role: roles/resourcemanager.organizationAdmin - members: - user:eve@example.com role: roles/resourcemanager.organizationViewer condition: title: expirable access description: Does not grant access after Sep 2020 expression: request.time < timestamp('2020-10-01T00:00:00.000Z') etag: BwWWja0YfJA= version: 3 ``` For a description of IAM and its features, see the [IAM documentation](https://cloud.google.com/iam/docs/).
+  "auditConfigs": [ # Specifies cloud audit logging configuration for this policy.
+    { # Specifies the audit configuration for a service. The configuration determines which permission types are logged, and what identities, if any, are exempted from logging. An AuditConfig must have one or more AuditLogConfigs. If there are AuditConfigs for both `allServices` and a specific service, the union of the two AuditConfigs is used for that service: the log_types specified in each AuditConfig are enabled, and the exempted_members in each AuditLogConfig are exempted. Example Policy with multiple AuditConfigs: { "audit_configs": [ { "service": "allServices", "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" }, { "log_type": "ADMIN_READ" } ] }, { "service": "sampleservice.googleapis.com", "audit_log_configs": [ { "log_type": "DATA_READ" }, { "log_type": "DATA_WRITE", "exempted_members": [ "user:aliya@example.com" ] } ] } ] } For sampleservice, this policy enables DATA_READ, DATA_WRITE and ADMIN_READ logging. It also exempts `jose@example.com` from DATA_READ logging, and `aliya@example.com` from DATA_WRITE logging.
+      "auditLogConfigs": [ # The configuration for logging of each type of permission.
+        { # Provides the configuration for logging a type of permissions. Example: { "audit_log_configs": [ { "log_type": "DATA_READ", "exempted_members": [ "user:jose@example.com" ] }, { "log_type": "DATA_WRITE" } ] } This enables 'DATA_READ' and 'DATA_WRITE' logging, while exempting jose@example.com from DATA_READ logging.
+          "exemptedMembers": [ # Specifies the identities that do not cause logging for this type of permission. Follows the same format of Binding.members.
+            "A String",
+          ],
+          "logType": "A String", # The log type that this config enables.
+        },
+      ],
+      "service": "A String", # Specifies a service that will be enabled for audit logging. For example, `storage.googleapis.com`, `cloudsql.googleapis.com`. `allServices` is a special value that covers all services.
+    },
+  ],
+  "bindings": [ # Associates a list of `members`, or principals, with a `role`. Optionally, may specify a `condition` that determines how and when the `bindings` are applied. Each of the `bindings` must contain at least one principal. The `bindings` in a `Policy` can refer to up to 1,500 principals; up to 250 of these principals can be Google groups. Each occurrence of a principal counts towards these limits. For example, if the `bindings` grant 50 different roles to `user:alice@example.com`, and not to any other principal, then you can add another 1,450 principals to the `bindings` in the `Policy`.
+    { # Associates `members`, or principals, with a `role`.
+      "condition": { # Represents a textual expression in the Common Expression Language (CEL) syntax. CEL is a C-like expression language. The syntax and semantics of CEL are documented at https://github.com/google/cel-spec. Example (Comparison): title: "Summary size limit" description: "Determines if a summary is less than 100 chars" expression: "document.summary.size() < 100" Example (Equality): title: "Requestor is owner" description: "Determines if requestor is the document owner" expression: "document.owner == request.auth.claims.email" Example (Logic): title: "Public documents" description: "Determine whether the document should be publicly visible" expression: "document.type != 'private' && document.type != 'internal'" Example (Data Manipulation): title: "Notification string" description: "Create a notification string with a timestamp." expression: "'New message received at ' + string(document.create_time)" The exact variables and functions that may be referenced within an expression are determined by the service that evaluates it. See the service documentation for additional information. # The condition that is associated with this binding. If the condition evaluates to `true`, then this binding applies to the current request. If the condition evaluates to `false`, then this binding does not apply to the current request. However, a different role binding might grant the same role to one or more of the principals in this binding. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+        "description": "A String", # Optional. Description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
+        "expression": "A String", # Textual representation of an expression in Common Expression Language syntax.
+        "location": "A String", # Optional. String indicating the location of the expression for error reporting, e.g. a file name and a position in the file.
+        "title": "A String", # Optional. Title for the expression, i.e. a short string describing its purpose. This can be used e.g. in UIs which allow to enter the expression.
+      },
+      "members": [ # Specifies the principals requesting access for a Google Cloud resource. `members` can have the following values: * `allUsers`: A special identifier that represents anyone who is on the internet; with or without a Google account. * `allAuthenticatedUsers`: A special identifier that represents anyone who is authenticated with a Google account or a service account. Does not include identities that come from external identity providers (IdPs) through identity federation. * `user:{emailid}`: An email address that represents a specific Google account. For example, `alice@example.com` . * `serviceAccount:{emailid}`: An email address that represents a Google service account. For example, `my-other-app@appspot.gserviceaccount.com`. * `serviceAccount:{projectid}.svc.id.goog[{namespace}/{kubernetes-sa}]`: An identifier for a [Kubernetes service account](https://cloud.google.com/kubernetes-engine/docs/how-to/kubernetes-service-accounts). For example, `my-project.svc.id.goog[my-namespace/my-kubernetes-sa]`. * `group:{emailid}`: An email address that represents a Google group. For example, `admins@example.com`. * `domain:{domain}`: The G Suite domain (primary) that represents all the users of that domain. For example, `google.com` or `example.com`. * `principal://iam.googleapis.com/locations/global/workforcePools/{pool_id}/subject/{subject_attribute_value}`: A single identity in a workforce identity pool. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/group/{group_id}`: All workforce identities in a group. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/attribute.{attribute_name}/{attribute_value}`: All workforce identities with a specific attribute value. * `principalSet://iam.googleapis.com/locations/global/workforcePools/{pool_id}/*`: All identities in a workforce identity pool. * `principal://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/subject/{subject_attribute_value}`: A single identity in a workload identity pool. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/group/{group_id}`: A workload identity pool group. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/attribute.{attribute_name}/{attribute_value}`: All identities in a workload identity pool with a certain attribute. * `principalSet://iam.googleapis.com/projects/{project_number}/locations/global/workloadIdentityPools/{pool_id}/*`: All identities in a workload identity pool. * `deleted:user:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a user that has been recently deleted. For example, `alice@example.com?uid=123456789012345678901`. If the user is recovered, this value reverts to `user:{emailid}` and the recovered user retains the role in the binding. * `deleted:serviceAccount:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a service account that has been recently deleted. For example, `my-other-app@appspot.gserviceaccount.com?uid=123456789012345678901`. If the service account is undeleted, this value reverts to `serviceAccount:{emailid}` and the undeleted service account retains the role in the binding. * `deleted:group:{emailid}?uid={uniqueid}`: An email address (plus unique identifier) representing a Google group that has been recently deleted. For example, `admins@example.com?uid=123456789012345678901`. If the group is recovered, this value reverts to `group:{emailid}` and the recovered group retains the role in the binding. * `deleted:principal://iam.googleapis.com/locations/global/workforcePools/{pool_id}/subject/{subject_attribute_value}`: Deleted single identity in a workforce identity pool. For example, `deleted:principal://iam.googleapis.com/locations/global/workforcePools/my-pool-id/subject/my-subject-attribute-value`.
+        "A String",
+      ],
+      "role": "A String", # Role that is assigned to the list of `members`, or principals. For example, `roles/viewer`, `roles/editor`, or `roles/owner`. For an overview of the IAM roles and permissions, see the [IAM documentation](https://cloud.google.com/iam/docs/roles-overview). For a list of the available pre-defined roles, see [here](https://cloud.google.com/iam/docs/understanding-roles).
+    },
+  ],
+  "etag": "A String", # `etag` is used for optimistic concurrency control as a way to help prevent simultaneous updates of a policy from overwriting each other. It is strongly suggested that systems make use of the `etag` in the read-modify-write cycle to perform policy updates in order to avoid race conditions: An `etag` is returned in the response to `getIamPolicy`, and systems are expected to put that etag in the request to `setIamPolicy` to ensure that their change will be applied to the same version of the policy. **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost.
+  "version": 42, # Specifies the format of the policy. Valid values are `0`, `1`, and `3`. Requests that specify an invalid value are rejected. Any operation that affects conditional role bindings must specify version `3`. This requirement applies to the following operations: * Getting a policy that includes a conditional role binding * Adding a conditional role binding to a policy * Changing a conditional role binding in a policy * Removing any role binding, with or without a condition, from a policy that includes conditions **Important:** If you use IAM Conditions, you must include the `etag` field whenever you call `setIamPolicy`. If you omit this field, then IAM allows you to overwrite a version `3` policy with a version `1` policy, and all of the conditions in the version `3` policy are lost. If a policy does not include any conditions, operations on that policy may specify any valid version or leave the field unset. To learn which resources support conditions in their IAM policies, see the [IAM documentation](https://cloud.google.com/iam/help/conditions/resource-policies).
+}
+
+
setPeering(securityGateway, body=None, x__xgafv=None)
This is a custom method to allow customers to create a peering connections between Google network and customer networks. This is enabled only for the allowlisted customers.
diff --git a/docs/dyn/chat_v1.spaces.html b/docs/dyn/chat_v1.spaces.html
index 51c9e048d5f..f0ead9c4a45 100644
--- a/docs/dyn/chat_v1.spaces.html
+++ b/docs/dyn/chat_v1.spaces.html
@@ -906,17 +906,17 @@ 

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, diff --git a/docs/dyn/chat_v1.spaces.members.html b/docs/dyn/chat_v1.spaces.members.html index ff61d2cc797..4b38be6fa9f 100644 --- a/docs/dyn/chat_v1.spaces.members.html +++ b/docs/dyn/chat_v1.spaces.members.html @@ -113,22 +113,22 @@

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. } - useAdminAccess: boolean, When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Creating app memberships or creating memberships for users outside the administrator's Google Workspace organization isn't supported using admin access. + useAdminAccess: boolean, Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Creating app memberships or creating memberships for users outside the administrator's Google Workspace organization isn't supported using admin access. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -140,17 +140,17 @@

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }
@@ -162,7 +162,7 @@

Method Details

Args: name: string, Required. Resource name of the membership to delete. Chat apps can delete human users' or their own memberships. Chat apps can't delete other apps' memberships. When deleting a human membership, requires the `chat.memberships` scope and `spaces/{space}/members/{member}` format. You can use the email as an alias for `{member}`. For example, `spaces/{space}/members/example@gmail.com` where `example@gmail.com` is the email of the Google Chat user. When deleting an app membership, requires the `chat.memberships.app` scope and `spaces/{space}/members/app` format. Format: `spaces/{space}/members/{member}` or `spaces/{space}/members/app`. (required) - useAdminAccess: boolean, When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Deleting app memberships in a space isn't supported using admin access. + useAdminAccess: boolean, Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Deleting app memberships in a space isn't supported using admin access. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -174,17 +174,17 @@

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. } @@ -196,7 +196,7 @@

Method Details

Args: name: string, Required. Resource name of the membership to retrieve. To get the app's own membership [by using user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), you can optionally use `spaces/{space}/members/app`. Format: `spaces/{space}/members/{member}` or `spaces/{space}/members/app` You can use the user's email as an alias for `{member}`. For example, `spaces/{space}/members/example@gmail.com` where `example@gmail.com` is the email of the Google Chat user. (required) - useAdminAccess: boolean, When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` or `chat.admin.memberships.readonly` [OAuth 2.0 scopes](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Getting app memberships in a space isn't supported when using admin access. + useAdminAccess: boolean, Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` or `chat.admin.memberships.readonly` [OAuth 2.0 scopes](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Getting app memberships in a space isn't supported when using admin access. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -208,17 +208,17 @@

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. } @@ -235,7 +235,7 @@

Method Details

pageToken: string, Optional. A page token, received from a previous call to list memberships. Provide this parameter to retrieve the subsequent page. When paginating, all other parameters provided should match the call that provided the page token. Passing different values to the other parameters might lead to unexpected results. showGroups: boolean, Optional. When `true`, also returns memberships associated with a Google Group, in addition to other types of memberships. If a filter is set, Google Group memberships that don't match the filter criteria aren't returned. showInvited: boolean, Optional. When `true`, also returns memberships associated with invited members, in addition to other types of memberships. If a filter is set, invited memberships that don't match the filter criteria aren't returned. Currently requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). - useAdminAccess: boolean, When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires either the `chat.admin.memberships.readonly` or `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Listing app memberships in a space isn't supported when using admin access. + useAdminAccess: boolean, Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires either the `chat.admin.memberships.readonly` or `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Listing app memberships in a space isn't supported when using admin access. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -249,17 +249,17 @@

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -287,30 +287,30 @@

Method Details

Updates a membership. For an example, see [Update a user's membership in a space](https://developers.google.com/workspace/chat/update-members). Supports the following types of [authentication](https://developers.google.com/workspace/chat/authenticate-authorize): - [App authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-app) with [administrator approval](https://support.google.com/a?p=chat-app-auth) in [Developer Preview](https://developers.google.com/workspace/preview) - [User authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user) You can authenticate and authorize this method with administrator privileges by setting the `use_admin_access` field in the request.
 
 Args:
-  name: string, Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` (required)
+  name: string, Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` (required)
   body: object, The request body.
     The object takes the form of:
 
 { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space.
   "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces.
   "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces.
-  "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user).
+  "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user).
     "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group}
   },
-  "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`.
+  "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`.
     "displayName": "A String", # Output only. The user's display name.
     "domainId": "A String", # Unique identifier of the user's Google Workspace domain.
     "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible.
     "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API.
     "type": "A String", # User type.
   },
-  "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`
+  "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`
   "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`.
   "state": "A String", # Output only. State of the membership.
 }
 
   updateMask: string, Required. The field paths to update. Separate multiple values with commas or use `*` to update all field paths. Currently supported field paths: - `role`
-  useAdminAccess: boolean, When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes).
+  useAdminAccess: boolean, Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes).
   x__xgafv: string, V1 error format.
     Allowed values
       1 - v1 error format
@@ -322,17 +322,17 @@ 

Method Details

{ # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }
diff --git a/docs/dyn/chat_v1.spaces.spaceEvents.html b/docs/dyn/chat_v1.spaces.spaceEvents.html index 9db73633a6e..7aa98924e42 100644 --- a/docs/dyn/chat_v1.spaces.spaceEvents.html +++ b/docs/dyn/chat_v1.spaces.spaceEvents.html @@ -115,17 +115,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The new membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -138,17 +138,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The deleted membership. Only the `name` and `state` fields are populated. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -161,17 +161,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The updated membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -182,17 +182,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The new membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -201,17 +201,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The deleted membership. Only the `name` and `state` fields are populated. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -220,17 +220,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The updated membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12419,7 +12419,7 @@

Method Details

parent: string, Required. Resource name of the [Google Chat space](https://developers.google.com/workspace/chat/api/reference/rest/v1/spaces) where the events occurred. Format: `spaces/{space}`. (required) filter: string, Required. A query filter. You must specify at least one event type (`event_type`) using the has `:` operator. To filter by multiple event types, use the `OR` operator. Omit batch event types in your filter. The request automatically returns any related batch events. For example, if you filter by new reactions (`google.workspace.chat.reaction.v1.created`), the server also returns batch new reactions events (`google.workspace.chat.reaction.v1.batchCreated`). For a list of supported event types, see the [`SpaceEvents` reference documentation](https://developers.google.com/workspace/chat/api/reference/rest/v1/spaces.spaceEvents#SpaceEvent.FIELDS.event_type). Optionally, you can also filter by start time (`start_time`) and end time (`end_time`): * `start_time`: Exclusive timestamp from which to start listing space events. You can list events that occurred up to 28 days ago. If unspecified, lists space events from the past 28 days. * `end_time`: Inclusive timestamp until which space events are listed. If unspecified, lists events up to the time of the request. To specify a start or end time, use the equals `=` operator and format in [RFC-3339](https://www.rfc-editor.org/rfc/rfc3339). To filter by both `start_time` and `end_time`, use the `AND` operator. For example, the following queries are valid: ``` start_time="2023-08-23T19:20:33+00:00" AND end_time="2023-08-23T19:21:54+00:00" ``` ``` start_time="2023-08-23T19:20:33+00:00" AND (event_types:"google.workspace.chat.space.v1.updated" OR event_types:"google.workspace.chat.message.v1.created") ``` The following queries are invalid: ``` start_time="2023-08-23T19:20:33+00:00" OR end_time="2023-08-23T19:21:54+00:00" ``` ``` event_types:"google.workspace.chat.space.v1.updated" AND event_types:"google.workspace.chat.message.v1.created" ``` Invalid queries are rejected by the server with an `INVALID_ARGUMENT` error. pageSize: integer, Optional. The maximum number of space events returned. The service might return fewer than this value. Negative values return an `INVALID_ARGUMENT` error. - pageToken: string, A page token, received from a previous list space events call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to list space events must match the call that provided the page token. Passing different values to the other parameters might lead to unexpected results. + pageToken: string, Optional. A page token, received from a previous list space events call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to list space events must match the call that provided the page token. Passing different values to the other parameters might lead to unexpected results. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format @@ -12440,17 +12440,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The new membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12463,17 +12463,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The deleted membership. Only the `name` and `state` fields are populated. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12486,17 +12486,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The updated membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12507,17 +12507,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The new membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12526,17 +12526,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The deleted membership. Only the `name` and `state` fields are populated. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, @@ -12545,17 +12545,17 @@

Method Details

"membership": { # Represents a membership relation in Google Chat, such as whether a user or Chat app is invited to, part of, or absent from a space. # The updated membership. "createTime": "A String", # Optional. Immutable. The creation time of the membership, such as when a member joined or was invited to join a space. This field is output only, except when used to import historical memberships in import mode spaces. "deleteTime": "A String", # Optional. Immutable. The deletion time of the membership, such as when a member left or was removed from a space. This field is output only, except when used to import historical memberships in import mode spaces. - "groupMember": { # A Google Group in Google Chat. # The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). + "groupMember": { # A Google Group in Google Chat. # Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user). "name": "A String", # Resource name for a Google Group. Represents a [group](https://cloud.google.com/identity/docs/reference/rest/v1/groups) in Cloud Identity Groups API. Format: groups/{group} }, - "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. + "member": { # A user in Google Chat. When returned as an output from a request, if your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output for a `User` resource only populates the user's `name` and `type`. # Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`. "displayName": "A String", # Output only. The user's display name. "domainId": "A String", # Unique identifier of the user's Google Workspace domain. "isAnonymous": True or False, # Output only. When `true`, the user is deleted or their profile is not visible. "name": "A String", # Resource name for a Google Chat user. Format: `users/{user}`. `users/app` can be used as an alias for the calling app bot user. For human users, `{user}` is the same user identifier as: - the `id` for the [Person](https://developers.google.com/people/api/rest/v1/people) in the People API. For example, `users/123456789` in Chat API represents the same person as the `123456789` Person profile ID in People API. - the `id` for a [user](https://developers.google.com/admin-sdk/directory/reference/rest/v1/users) in the Admin SDK Directory API. - the user's email address can be used as an alias for `{user}` in API requests. For example, if the People API Person profile ID for `user@example.com` is `123456789`, you can use `users/user@example.com` as an alias to reference `users/123456789`. Only the canonical resource name (for example `users/123456789`) will be returned from the API. "type": "A String", # User type. }, - "name": "A String", # Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` + "name": "A String", # Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}` "role": "A String", # Optional. User's role within a Chat space, which determines their permitted actions in the space. This field can only be used as input in `UpdateMembership`. "state": "A String", # Output only. State of the membership. }, diff --git a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automationRuns.html b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automationRuns.html index a9a4f7b1448..8393ce80c3b 100644 --- a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automationRuns.html +++ b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automationRuns.html @@ -163,6 +163,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -179,6 +188,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -194,6 +212,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -216,6 +243,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. @@ -285,7 +337,12 @@

Method Details

"serviceAccount": "A String", # Output only. Email address of the user-managed IAM service account that performs the operations against Cloud Deploy resources. "state": "A String", # Output only. Current state of the `AutomationRun`. "stateDescription": "A String", # Output only. Explains the current state of the `AutomationRun`. Present only when an explanation is needed. - "targetId": "A String", # Output only. The ID of the target that represents the promotion stage that initiates the `AutomationRun`. The value of this field is the last segment of a target name. + "targetId": "A String", # Output only. The ID of the source target that initiates the `AutomationRun`. The value of this field is the last segment of a target name. + "timedPromoteReleaseOperation": { # Contains the information of an automated timed promote-release operation. # Output only. Promotes a release to a specified 'Target' as defined in a Timed Promote Release rule. + "phase": "A String", # Output only. The starting phase of the rollout created by this operation. + "release": "A String", # Output only. The name of the release to be promoted. + "targetId": "A String", # Output only. The ID of the target that represents the promotion stage to which the release will be promoted. The value of this field is the last segment of a target name. + }, "updateTime": "A String", # Output only. Time at which the automationRun was updated. "waitUntilTime": "A String", # Output only. Earliest time the `AutomationRun` will attempt to resume. Wait-time is configured by `wait` in automation rule. } @@ -341,6 +398,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -357,6 +423,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -372,6 +447,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -394,6 +478,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. @@ -463,7 +572,12 @@

Method Details

"serviceAccount": "A String", # Output only. Email address of the user-managed IAM service account that performs the operations against Cloud Deploy resources. "state": "A String", # Output only. Current state of the `AutomationRun`. "stateDescription": "A String", # Output only. Explains the current state of the `AutomationRun`. Present only when an explanation is needed. - "targetId": "A String", # Output only. The ID of the target that represents the promotion stage that initiates the `AutomationRun`. The value of this field is the last segment of a target name. + "targetId": "A String", # Output only. The ID of the source target that initiates the `AutomationRun`. The value of this field is the last segment of a target name. + "timedPromoteReleaseOperation": { # Contains the information of an automated timed promote-release operation. # Output only. Promotes a release to a specified 'Target' as defined in a Timed Promote Release rule. + "phase": "A String", # Output only. The starting phase of the rollout created by this operation. + "release": "A String", # Output only. The name of the release to be promoted. + "targetId": "A String", # Output only. The ID of the target that represents the promotion stage to which the release will be promoted. The value of this field is the last segment of a target name. + }, "updateTime": "A String", # Output only. Time at which the automationRun was updated. "waitUntilTime": "A String", # Output only. Earliest time the `AutomationRun` will attempt to resume. Wait-time is configured by `wait` in automation rule. }, diff --git a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automations.html b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automations.html index d96a70d0f27..0fbc11a6dec 100644 --- a/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automations.html +++ b/docs/dyn/clouddeploy_v1.projects.locations.deliveryPipelines.automations.html @@ -132,6 +132,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -148,6 +157,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -163,6 +181,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -185,6 +212,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. @@ -310,6 +362,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -326,6 +387,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -341,6 +411,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -363,6 +442,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. @@ -424,6 +528,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -440,6 +553,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -455,6 +577,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -477,6 +608,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. @@ -547,6 +703,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "sourcePhases": [ # Optional. Proceeds only after phase name matched any one in the list. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -563,6 +728,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this operation. Default to the first phase. "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence @@ -578,6 +752,15 @@

Method Details

"status": True or False, # True if there aren't any missing Targets. "updateTime": "A String", # Last time the condition was updated. }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, }, "id": "A String", # Required. ID of the rule. This id must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. "jobs": [ # Optional. Jobs to repair. Proceeds only after job name matched any one in the list, or for all jobs if unspecified or empty. The phase that includes the job must match the phase ID specified in `source_phase`. This value must consist of lower-case letters, numbers, and hyphens, start with a letter and end with a letter or a number, and have a max length of 63 characters. In other words, it must match the following regex: `^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$`. @@ -600,6 +783,31 @@

Method Details

}, ], }, + "timedPromoteReleaseRule": { # The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. # Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule. + "condition": { # `AutomationRuleCondition` contains conditions relevant to an `Automation` rule. # Output only. Information around the state of the Automation rule. + "targetsPresentCondition": { # `TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist. # Optional. Details around targets enumerated in the rule. + "missingTargets": [ # The list of Target names that do not exist. For example, `projects/{project_id}/locations/{location_name}/targets/{target_name}`. + "A String", + ], + "status": True or False, # True if there aren't any missing Targets. + "updateTime": "A String", # Last time the condition was updated. + }, + "timedPromoteReleaseCondition": { # `TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined. # Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined. + "nextPromotionTime": "A String", # Output only. When the next scheduled promotion(s) will occur. + "targetsList": [ # Output only. A list of targets involved in the upcoming timed promotion(s). + { # The targets involved in a single timed promotion. + "destinationTargetId": "A String", # Optional. The destination target ID. + "sourceTargetId": "A String", # Optional. The source target ID. + }, + ], + }, + }, + "destinationPhase": "A String", # Optional. The starting phase of the rollout created by this rule. Default to the first phase. + "destinationTargetId": "A String", # Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * "@next", the next target in the promotion sequence + "id": "A String", # Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`. + "schedule": "A String", # Required. Schedule in crontab format. e.g. "0 9 * * 1" for every Monday at 9am. + "timeZone": "A String", # Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York). + }, }, ], "selector": { # AutomationResourceSelector contains the information to select the resources to which an Automation is going to be applied. # Required. Selected resources to which the automation will be applied. diff --git a/docs/dyn/cloudfunctions_v2.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2.projects.locations.functions.html index 5483b3dad2e..60d9e24eef9 100644 --- a/docs/dyn/cloudfunctions_v2.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2.projects.locations.functions.html @@ -352,7 +352,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -737,7 +737,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1053,7 +1053,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1330,7 +1330,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, diff --git a/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html index e51c9642232..31cbcbc22fc 100644 --- a/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html @@ -352,7 +352,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -737,7 +737,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1053,7 +1053,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1330,7 +1330,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, diff --git a/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html index 787334e1708..d75c619d68e 100644 --- a/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html @@ -352,7 +352,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -737,7 +737,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1053,7 +1053,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, @@ -1330,7 +1330,7 @@

Method Details

}, ], "updateTime": "A String", # Output only. The last update timestamp of a Cloud Function. - "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration. # Output only. UpgradeInfo for this Cloud Function + "upgradeInfo": { # Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach. # Output only. UpgradeInfo for this Cloud Function "buildConfig": { # Describes the Build step of the function that builds a container from the given source. # Describes the Build step of the function that builds a container to prepare for 2nd gen upgrade. "automaticUpdatePolicy": { # Security patches are applied automatically to the runtime without requiring the function to be redeployed. }, diff --git a/docs/dyn/composer_v1.projects.locations.environments.html b/docs/dyn/composer_v1.projects.locations.environments.html index 2a6e2251d38..606f127962f 100644 --- a/docs/dyn/composer_v1.projects.locations.environments.html +++ b/docs/dyn/composer_v1.projects.locations.environments.html @@ -192,7 +192,7 @@

Method Details

The object takes the form of: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -228,7 +228,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -253,7 +253,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -281,7 +281,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -291,7 +291,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -346,7 +346,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -531,7 +531,7 @@

Method Details

An object of the form: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -567,7 +567,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -592,7 +592,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -620,7 +620,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -630,7 +630,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -685,7 +685,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -716,7 +716,7 @@

Method Details

{ # The environments in a project and location. "environments": [ # The list of environments returned by a ListEnvironmentsRequest. { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -752,7 +752,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -777,7 +777,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -805,7 +805,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -815,7 +815,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -870,7 +870,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -955,7 +955,7 @@

Method Details

The object takes the form of: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -991,7 +991,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -1016,7 +1016,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -1044,7 +1044,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -1054,7 +1054,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -1109,7 +1109,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. diff --git a/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsConfigMaps.html b/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsConfigMaps.html index 90fbc587502..f1f974e0c8b 100644 --- a/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsConfigMaps.html +++ b/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsConfigMaps.html @@ -111,7 +111,7 @@

Method Details

The object takes the form of: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -126,7 +126,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -166,7 +166,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -193,7 +193,7 @@

Method Details

"nextPageToken": "A String", # The page token used to query for the next page if one exists. "userWorkloadsConfigMaps": [ # The list of ConfigMaps returned by a ListUserWorkloadsConfigMapsRequest. { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -226,7 +226,7 @@

Method Details

The object takes the form of: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -241,7 +241,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" diff --git a/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsSecrets.html b/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsSecrets.html index 2803f18fecc..27d9eab2dc4 100644 --- a/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsSecrets.html +++ b/docs/dyn/composer_v1.projects.locations.environments.userWorkloadsSecrets.html @@ -111,7 +111,7 @@

Method Details

The object takes the form of: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -126,7 +126,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -166,7 +166,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -193,7 +193,7 @@

Method Details

"nextPageToken": "A String", # The page token used to query for the next page if one exists. "userWorkloadsSecrets": [ # The list of Secrets returned by a ListUserWorkloadsSecretsRequest. { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -226,7 +226,7 @@

Method Details

The object takes the form of: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -241,7 +241,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" diff --git a/docs/dyn/composer_v1beta1.projects.locations.environments.html b/docs/dyn/composer_v1beta1.projects.locations.environments.html index 82dd8d0b9b1..73f61bf29bb 100644 --- a/docs/dyn/composer_v1beta1.projects.locations.environments.html +++ b/docs/dyn/composer_v1beta1.projects.locations.environments.html @@ -195,7 +195,7 @@

Method Details

The object takes the form of: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -232,7 +232,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -258,7 +258,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -286,7 +286,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -296,7 +296,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -351,7 +351,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -536,7 +536,7 @@

Method Details

An object of the form: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -573,7 +573,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -599,7 +599,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -627,7 +627,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -637,7 +637,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -692,7 +692,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -723,7 +723,7 @@

Method Details

{ # The environments in a project and location. "environments": [ # The list of environments returned by a ListEnvironmentsRequest. { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -760,7 +760,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -786,7 +786,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -814,7 +814,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -824,7 +824,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -879,7 +879,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. @@ -964,7 +964,7 @@

Method Details

The object takes the form of: { # An environment for running orchestration tasks. - "config": { # Configuration information for an environment. # Configuration parameters for this environment. + "config": { # Configuration information for an environment. # Optional. Configuration parameters for this environment. "airflowByoidUri": "A String", # Output only. The 'bring your own identity' variant of the URI of the Apache Airflow Web UI hosted within this environment, to be accessed with external identities using workforce identity federation (see [Access environments with workforce identity federation](/composer/docs/composer-2/access-environments-with-workforce-identity-federation)). "airflowUri": "A String", # Output only. The URI of the Apache Airflow Web UI hosted within this environment (see [Airflow web interface](/composer/docs/how-to/accessing/airflow-web-interface)). "dagGcsPrefix": "A String", # Output only. The Cloud Storage prefix of the DAGs for this environment. Although Cloud Storage objects reside in a flat namespace, a hierarchical file tree can be simulated using "/"-delimited object name prefixes. DAG objects for this environment reside in a simulated directory with the given prefix. @@ -1001,7 +1001,7 @@

Method Details

], "enabled": True or False, # Whether or not master authorized networks feature is enabled. }, - "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # The configuration used for the Kubernetes Engine cluster. + "nodeConfig": { # The configuration information for the Kubernetes Engine nodes running the Apache Airflow software. # Optional. The configuration used for the Kubernetes Engine cluster. "composerInternalIpv4CidrBlock": "A String", # Optional. The IP range in CIDR notation to use internally by Cloud Composer. IP addresses are not reserved - and the same range can be used by multiple Cloud Composer environments. In case of overlap, IPs from this range will not be accessible in the user's VPC network. Cannot be updated. If not specified, the default value of '100.64.128.0/20' is used. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "composerNetworkAttachment": "A String", # Optional. Network Attachment that Cloud Composer environment is connected to, which provides connectivity with a user's VPC network. Takes precedence over network and subnetwork settings. If not provided, but network and subnetwork are defined during environment, it will be provisioned. If not provided and network and subnetwork are also empty, then connectivity to user's VPC network is disabled. Network attachment must be provided in format projects/{project}/regions/{region}/networkAttachments/{networkAttachment}. This field is supported for Cloud Composer environments in versions composer-3.*.*-airflow-*.*.* and newer. "diskSizeGb": 42, # Optional. The disk size in GB used for node VMs. Minimum size is 30GB. If unspecified, defaults to 100GB. Cannot be updated. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. @@ -1027,7 +1027,7 @@

Method Details

], }, "nodeCount": 42, # The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*. - "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # The configuration used for the Private IP Cloud Composer environment. + "privateEnvironmentConfig": { # The configuration information for configuring a Private IP Cloud Composer environment. # Optional. The configuration used for the Private IP Cloud Composer environment. "cloudComposerConnectionSubnetwork": "A String", # Optional. When specified, the environment will use Private Service Connect instead of VPC peerings to connect to Cloud SQL in the Tenant Project, and the PSC endpoint in the Customer Project will use an IP address from this subnetwork. "cloudComposerNetworkIpv4CidrBlock": "A String", # Optional. The CIDR block from which IP range for Cloud Composer Network in tenant project will be reserved. Needs to be disjoint from private_cluster_config.master_ipv4_cidr_block and cloud_sql_ipv4_cidr_block. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. "cloudComposerNetworkIpv4ReservedRange": "A String", # Output only. The IP range reserved for the tenant project's Cloud Composer network. This field is supported for Cloud Composer environments in versions composer-2.*.*-airflow-*.*.* and newer. @@ -1055,7 +1055,7 @@

Method Details

}, }, "resilienceMode": "A String", # Optional. Resilience mode of the Cloud Composer Environment. This field is supported for Cloud Composer environments in versions composer-2.2.0-airflow-*.*.* and newer. - "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # The configuration settings for software inside the environment. + "softwareConfig": { # Specifies the selection and configuration of software inside the environment. # Optional. The configuration settings for software inside the environment. "airflowConfigOverrides": { # Optional. Apache Airflow configuration properties to override. Property keys contain the section and property names, separated by a hyphen, for example "core-dags_are_paused_at_creation". Section names must not contain hyphens ("-"), opening square brackets ("["), or closing square brackets ("]"). The property name must not be empty and must not contain an equals sign ("=") or semicolon (";"). Section and property names must not contain a period ("."). Apache Airflow configuration property names must be written in [snake_case](https://en.wikipedia.org/wiki/Snake_case). Property values can contain any character, and can be written in any lower/upper case format. Certain Apache Airflow configuration property values are [blocked](/composer/docs/concepts/airflow-configurations), and cannot be overridden. "a_key": "A String", }, @@ -1065,7 +1065,7 @@

Method Details

"envVariables": { # Optional. Additional environment variables to provide to the Apache Airflow scheduler, worker, and webserver processes. Environment variable names must match the regular expression `a-zA-Z_*`. They cannot specify Apache Airflow software configuration overrides (they cannot match the regular expression `AIRFLOW__[A-Z0-9_]+__[A-Z0-9_]+`), and they cannot match any of the following reserved names: * `AIRFLOW_HOME` * `C_FORCE_ROOT` * `CONTAINER_NAME` * `DAGS_FOLDER` * `GCP_PROJECT` * `GCS_BUCKET` * `GKE_CLUSTER_NAME` * `SQL_DATABASE` * `SQL_INSTANCE` * `SQL_PASSWORD` * `SQL_PROJECT` * `SQL_REGION` * `SQL_USER` "a_key": "A String", }, - "imageVersion": "A String", # The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). + "imageVersion": "A String", # Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\.[0-9]+\.[0-9]+(-preview\.[0-9]+)?)?|latest)-airflow-([0-9]+(\.[0-9]+(\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview). "pypiPackages": { # Optional. Custom Python Package Index (PyPI) packages to be installed in the environment. Keys refer to the lowercase package name such as "numpy" and values are the lowercase extras and version specifier such as "==1.12.0", "[devel,gcp_api]", or "[devel]>=1.8.2, <1.9.2". To specify a package without pinning it to a version specifier, use the empty string as the value. "a_key": "A String", }, @@ -1120,7 +1120,7 @@

Method Details

"labels": { # Optional. User-defined labels for this environment. The labels map can contain no more than 64 entries. Entries of the labels map are UTF8 strings that comply with the following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and values are additionally constrained to be <= 128 bytes in size. "a_key": "A String", }, - "name": "A String", # The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. + "name": "A String", # Identifier. The resource name of the environment, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen. "satisfiesPzi": True or False, # Output only. Reserved for future use. "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # The current state of the environment. diff --git a/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsConfigMaps.html b/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsConfigMaps.html index 7f4b0cd433c..3cc50446302 100644 --- a/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsConfigMaps.html +++ b/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsConfigMaps.html @@ -111,7 +111,7 @@

Method Details

The object takes the form of: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -126,7 +126,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -166,7 +166,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -193,7 +193,7 @@

Method Details

"nextPageToken": "A String", # The page token used to query for the next page if one exists. "userWorkloadsConfigMaps": [ # The list of ConfigMaps returned by a ListUserWorkloadsConfigMapsRequest. { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -226,7 +226,7 @@

Method Details

The object takes the form of: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" @@ -241,7 +241,7 @@

Method Details

An object of the form: { # User workloads ConfigMap used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ + "data": { # Optional. The "data" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { "example_key": "example_value", "another_key": "another_value" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the ConfigMap, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsConfigMaps/{userWorkloadsConfigMapId}" diff --git a/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsSecrets.html b/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsSecrets.html index 6a0708c9845..111be48162f 100644 --- a/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsSecrets.html +++ b/docs/dyn/composer_v1beta1.projects.locations.environments.userWorkloadsSecrets.html @@ -111,7 +111,7 @@

Method Details

The object takes the form of: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -126,7 +126,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -166,7 +166,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -193,7 +193,7 @@

Method Details

"nextPageToken": "A String", # The page token used to query for the next page if one exists. "userWorkloadsSecrets": [ # The list of Secrets returned by a ListUserWorkloadsSecretsRequest. { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -226,7 +226,7 @@

Method Details

The object takes the form of: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" @@ -241,7 +241,7 @@

Method Details

An object of the form: { # User workloads Secret used by Airflow tasks that run with Kubernetes executor or KubernetesPodOperator. - "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ + "data": { # Optional. The "data" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { "example": "ZXhhbXBsZV92YWx1ZQ==", "another-example": "YW5vdGhlcl9leGFtcGxlX3ZhbHVl" } "a_key": "A String", }, "name": "A String", # Identifier. The resource name of the Secret, in the form: "projects/{projectId}/locations/{locationId}/environments/{environmentId}/userWorkloadsSecrets/{userWorkloadsSecretId}" diff --git a/docs/dyn/compute_v1.backendBuckets.html b/docs/dyn/compute_v1.backendBuckets.html index 97f6e9b8af6..fa912fa31fc 100644 --- a/docs/dyn/compute_v1.backendBuckets.html +++ b/docs/dyn/compute_v1.backendBuckets.html @@ -534,7 +534,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -646,7 +646,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -838,7 +838,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -931,7 +931,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1373,7 +1373,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. diff --git a/docs/dyn/compute_v1.backendServices.html b/docs/dyn/compute_v1.backendServices.html index 92d59248695..57c91016f6f 100644 --- a/docs/dyn/compute_v1.backendServices.html +++ b/docs/dyn/compute_v1.backendServices.html @@ -335,7 +335,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -857,7 +857,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1171,7 +1171,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1518,7 +1518,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1764,7 +1764,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -2026,7 +2026,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -2754,7 +2754,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. diff --git a/docs/dyn/compute_v1.instances.html b/docs/dyn/compute_v1.instances.html index 9edaf3ac877..babab7d717f 100644 --- a/docs/dyn/compute_v1.instances.html +++ b/docs/dyn/compute_v1.instances.html @@ -2507,7 +2507,7 @@

Method Details

zone: string, The name of the zone for this request. (required) instance: string, Name of the instance for this request. (required) port: integer, Specifies which COM or serial port to retrieve data from. - start: string, Specifies the starting byte position of the output to return. To start with the first byte of output to the specified port, omit this field or set it to `0`. If the output for that byte position is available, this field matches the `start` parameter sent with the request. If the amount of serial console output exceeds the size of the buffer (1 MB), the oldest output is discarded and is no longer available. If the requested start position refers to discarded output, the start position is adjusted to the oldest output still available, and the adjusted start position is returned as the `start` property value. You can also provide a negative start position, which translates to the most recent number of bytes written to the serial port. For example, -3 is interpreted as the most recent 3 bytes written to the serial console. + start: string, Specifies the starting byte position of the output to return. To start with the first byte of output to the specified port, omit this field or set it to `0`. If the output for that byte position is available, this field matches the `start` parameter sent with the request. If the amount of serial console output exceeds the size of the buffer (1 MB), the oldest output is discarded and is no longer available. If the requested start position refers to discarded output, the start position is adjusted to the oldest output still available, and the adjusted start position is returned as the `start` property value. You can also provide a negative start position, which translates to the most recent number of bytes written to the serial port. For example, -3 is interpreted as the most recent 3 bytes written to the serial console. Note that the negative start is bounded by the retained buffer size, and the returned serial console output will not exceed the max buffer size. x__xgafv: string, V1 error format. Allowed values 1 - v1 error format diff --git a/docs/dyn/compute_v1.networks.html b/docs/dyn/compute_v1.networks.html index 1f9bf7bf38b..9e3f7cb396e 100644 --- a/docs/dyn/compute_v1.networks.html +++ b/docs/dyn/compute_v1.networks.html @@ -439,6 +439,9 @@

Method Details

}, ], "routingConfig": { # A routing configuration attached to a network resource. The message includes the list of routers associated with the network, and a flag indicating the type of routing behavior to enforce network-wide. # The network-level routing configuration for this network. Used by Cloud Router to determine what type of network-wide routing behavior to enforce. + "bgpAlwaysCompareMed": True or False, # Enable comparison of Multi-Exit Discriminators (MED) across routes with different neighbor ASNs when using the STANDARD BGP best path selection algorithm. + "bgpBestPathSelectionMode": "A String", # The BGP best path selection algorithm to be employed within this network for dynamic routes learned by Cloud Routers. Can be LEGACY (default) or STANDARD. + "bgpInterRegionCost": "A String", # Allows to define a preferred approach for handling inter-region cost in the selection process when using the STANDARD BGP best path selection algorithm. Can be DEFAULT or ADD_COST_TO_MED. "routingMode": "A String", # The network-wide routing mode to use. If set to REGIONAL, this network's Cloud Routers will only advertise routes with subnets of this network in the same region as the router. If set to GLOBAL, this network's Cloud Routers will advertise routes with all subnets of this network, across regions. }, "selfLink": "A String", # [Output Only] Server-defined URL for the resource. @@ -642,6 +645,9 @@

Method Details

}, ], "routingConfig": { # A routing configuration attached to a network resource. The message includes the list of routers associated with the network, and a flag indicating the type of routing behavior to enforce network-wide. # The network-level routing configuration for this network. Used by Cloud Router to determine what type of network-wide routing behavior to enforce. + "bgpAlwaysCompareMed": True or False, # Enable comparison of Multi-Exit Discriminators (MED) across routes with different neighbor ASNs when using the STANDARD BGP best path selection algorithm. + "bgpBestPathSelectionMode": "A String", # The BGP best path selection algorithm to be employed within this network for dynamic routes learned by Cloud Routers. Can be LEGACY (default) or STANDARD. + "bgpInterRegionCost": "A String", # Allows to define a preferred approach for handling inter-region cost in the selection process when using the STANDARD BGP best path selection algorithm. Can be DEFAULT or ADD_COST_TO_MED. "routingMode": "A String", # The network-wide routing mode to use. If set to REGIONAL, this network's Cloud Routers will only advertise routes with subnets of this network in the same region as the router. If set to GLOBAL, this network's Cloud Routers will advertise routes with all subnets of this network, across regions. }, "selfLink": "A String", # [Output Only] Server-defined URL for the resource. @@ -821,6 +827,9 @@

Method Details

}, ], "routingConfig": { # A routing configuration attached to a network resource. The message includes the list of routers associated with the network, and a flag indicating the type of routing behavior to enforce network-wide. # The network-level routing configuration for this network. Used by Cloud Router to determine what type of network-wide routing behavior to enforce. + "bgpAlwaysCompareMed": True or False, # Enable comparison of Multi-Exit Discriminators (MED) across routes with different neighbor ASNs when using the STANDARD BGP best path selection algorithm. + "bgpBestPathSelectionMode": "A String", # The BGP best path selection algorithm to be employed within this network for dynamic routes learned by Cloud Routers. Can be LEGACY (default) or STANDARD. + "bgpInterRegionCost": "A String", # Allows to define a preferred approach for handling inter-region cost in the selection process when using the STANDARD BGP best path selection algorithm. Can be DEFAULT or ADD_COST_TO_MED. "routingMode": "A String", # The network-wide routing mode to use. If set to REGIONAL, this network's Cloud Routers will only advertise routes with subnets of this network in the same region as the router. If set to GLOBAL, this network's Cloud Routers will advertise routes with all subnets of this network, across regions. }, "selfLink": "A String", # [Output Only] Server-defined URL for the resource. @@ -968,6 +977,9 @@

Method Details

}, ], "routingConfig": { # A routing configuration attached to a network resource. The message includes the list of routers associated with the network, and a flag indicating the type of routing behavior to enforce network-wide. # The network-level routing configuration for this network. Used by Cloud Router to determine what type of network-wide routing behavior to enforce. + "bgpAlwaysCompareMed": True or False, # Enable comparison of Multi-Exit Discriminators (MED) across routes with different neighbor ASNs when using the STANDARD BGP best path selection algorithm. + "bgpBestPathSelectionMode": "A String", # The BGP best path selection algorithm to be employed within this network for dynamic routes learned by Cloud Routers. Can be LEGACY (default) or STANDARD. + "bgpInterRegionCost": "A String", # Allows to define a preferred approach for handling inter-region cost in the selection process when using the STANDARD BGP best path selection algorithm. Can be DEFAULT or ADD_COST_TO_MED. "routingMode": "A String", # The network-wide routing mode to use. If set to REGIONAL, this network's Cloud Routers will only advertise routes with subnets of this network in the same region as the router. If set to GLOBAL, this network's Cloud Routers will advertise routes with all subnets of this network, across regions. }, "selfLink": "A String", # [Output Only] Server-defined URL for the resource. diff --git a/docs/dyn/compute_v1.regionBackendServices.html b/docs/dyn/compute_v1.regionBackendServices.html index 1394a3537cb..85adbc80d20 100644 --- a/docs/dyn/compute_v1.regionBackendServices.html +++ b/docs/dyn/compute_v1.regionBackendServices.html @@ -308,7 +308,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -625,7 +625,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -973,7 +973,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1220,7 +1220,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -1483,7 +1483,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. @@ -2084,7 +2084,7 @@

Method Details

"A String", ], }, - "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. + "cacheMode": "A String", # Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any "private", "no-store" or "no-cache" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC. "clientTtl": 42, # Specifies a separate client (e.g. browser client) maximum TTL. This is used to clamp the max-age (or Expires) value sent to the client. With FORCE_CACHE_ALL, the lesser of client_ttl and default_ttl is used for the response max-age directive, along with a "public" directive. For cacheable content in CACHE_ALL_STATIC mode, client_ttl clamps the max-age from the origin (if specified), or else sets the response max-age directive to the lesser of the client_ttl and default_ttl, and also ensures a "public" cache-control directive is present. If a client TTL is not specified, a default value (1 hour) will be used. The maximum allowed value is 31,622,400s (1 year). "defaultTtl": 42, # Specifies the default TTL for cached content served by this origin for responses that do not have an existing valid TTL (max-age or s-max-age). Setting a TTL of "0" means "always revalidate". The value of defaultTTL cannot be set to a value greater than that of maxTTL, but can be equal. When the cacheMode is set to FORCE_CACHE_ALL, the defaultTTL will overwrite the TTL set in all responses. The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. "maxTtl": 42, # Specifies the maximum allowed TTL for cached content served by this origin. Cache directives that attempt to set a max-age or s-maxage higher than this, or an Expires header more than maxTTL seconds in the future will be capped at the value of maxTTL, as if it were the value of an s-maxage Cache-Control directive. Headers sent to the client will not be modified. Setting a TTL of "0" means "always revalidate". The maximum allowed value is 31,622,400s (1 year), noting that infrequently accessed objects may be evicted from the cache before the defined TTL. diff --git a/docs/dyn/compute_v1.routers.html b/docs/dyn/compute_v1.routers.html index d1f1271609a..104d453b163 100644 --- a/docs/dyn/compute_v1.routers.html +++ b/docs/dyn/compute_v1.routers.html @@ -811,8 +811,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. @@ -857,8 +860,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. @@ -905,8 +911,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. diff --git a/docs/dyn/compute_v1.routes.html b/docs/dyn/compute_v1.routes.html index 022bc846ef4..839546a85ed 100644 --- a/docs/dyn/compute_v1.routes.html +++ b/docs/dyn/compute_v1.routes.html @@ -257,8 +257,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. @@ -312,8 +315,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. @@ -496,8 +502,11 @@

Method Details

"nextHopHub": "A String", # [Output Only] The full resource name of the Network Connectivity Center hub that will handle matching packets. "nextHopIlb": "A String", # The URL to a forwarding rule of type loadBalancingScheme=INTERNAL that should handle matching packets or the IP address of the forwarding Rule. For example, the following are all valid URLs: - https://www.googleapis.com/compute/v1/projects/project/regions/region /forwardingRules/forwardingRule - regions/region/forwardingRules/forwardingRule If an IP address is provided, must specify an IPv4 address in dot-decimal notation or an IPv6 address in RFC 4291 format. For example, the following are all valid IP addresses: - 10.128.0.56 - 2001:db8::2d9:51:0:0 - 2001:db8:0:0:2d9:51:0:0 IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. "nextHopInstance": "A String", # The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/ + "nextHopInterRegionCost": 42, # [Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions. "nextHopIp": "A String", # The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address. + "nextHopMed": 42, # [Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network. "nextHopNetwork": "A String", # The URL of the local network if it should handle matching packets. + "nextHopOrigin": "A String", # [Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE. "nextHopPeering": "A String", # [Output Only] The network peering name that should handle matching packets, which should conform to RFC1035. "nextHopVpnTunnel": "A String", # The URL to a VpnTunnel that should handle matching packets. "priority": 42, # The priority of this route. Priority is used to break ties in cases where there is more than one matching route of equal prefix length. In cases where multiple routes have equal prefix length, the one with the lowest-numbered priority value wins. The default value is `1000`. The priority value must be from `0` to `65535`, inclusive. diff --git a/docs/dyn/compute_v1.subnetworks.html b/docs/dyn/compute_v1.subnetworks.html index 3db6b983cd7..0a2828adfdf 100644 --- a/docs/dyn/compute_v1.subnetworks.html +++ b/docs/dyn/compute_v1.subnetworks.html @@ -176,7 +176,7 @@

Method Details

"network": "A String", # The URL of the network to which this subnetwork belongs, provided by the client when initially creating the subnetwork. This field can be set only at resource creation time. "privateIpGoogleAccess": True or False, # Whether the VMs in this subnet can access Google services without assigned external IP addresses. This field can be both set at resource creation time and updated using setPrivateIpGoogleAccess. "privateIpv6GoogleAccess": "A String", # This field is for internal use. This field can be both set at resource creation time and updated using patch. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "region": "A String", # URL of the region where the Subnetwork resides. This field can be set only at resource creation time. "reservedInternalRange": "A String", # The URL of the reserved internal range. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. @@ -542,7 +542,7 @@

Method Details

"network": "A String", # The URL of the network to which this subnetwork belongs, provided by the client when initially creating the subnetwork. This field can be set only at resource creation time. "privateIpGoogleAccess": True or False, # Whether the VMs in this subnet can access Google services without assigned external IP addresses. This field can be both set at resource creation time and updated using setPrivateIpGoogleAccess. "privateIpv6GoogleAccess": "A String", # This field is for internal use. This field can be both set at resource creation time and updated using patch. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "region": "A String", # URL of the region where the Subnetwork resides. This field can be set only at resource creation time. "reservedInternalRange": "A String", # The URL of the reserved internal range. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. @@ -646,7 +646,7 @@

Method Details

"network": "A String", # The URL of the network to which this subnetwork belongs, provided by the client when initially creating the subnetwork. This field can be set only at resource creation time. "privateIpGoogleAccess": True or False, # Whether the VMs in this subnet can access Google services without assigned external IP addresses. This field can be both set at resource creation time and updated using setPrivateIpGoogleAccess. "privateIpv6GoogleAccess": "A String", # This field is for internal use. This field can be both set at resource creation time and updated using patch. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "region": "A String", # URL of the region where the Subnetwork resides. This field can be set only at resource creation time. "reservedInternalRange": "A String", # The URL of the reserved internal range. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. @@ -829,7 +829,7 @@

Method Details

"network": "A String", # The URL of the network to which this subnetwork belongs, provided by the client when initially creating the subnetwork. This field can be set only at resource creation time. "privateIpGoogleAccess": True or False, # Whether the VMs in this subnet can access Google services without assigned external IP addresses. This field can be both set at resource creation time and updated using setPrivateIpGoogleAccess. "privateIpv6GoogleAccess": "A String", # This field is for internal use. This field can be both set at resource creation time and updated using patch. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "region": "A String", # URL of the region where the Subnetwork resides. This field can be set only at resource creation time. "reservedInternalRange": "A String", # The URL of the reserved internal range. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. @@ -889,7 +889,7 @@

Method Details

"ipCidrRange": "A String", # The range of internal addresses that are owned by this subnetwork. "ipv6AccessType": "A String", # The access type of IPv6 address this subnet holds. It's immutable and can only be specified during creation or the first time the subnet is updated into IPV4_IPV6 dual stack. "network": "A String", # Network URL. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. "secondaryIpRanges": [ # Secondary IP ranges. { # Secondary IP range of a usable subnetwork. @@ -983,7 +983,7 @@

Method Details

"network": "A String", # The URL of the network to which this subnetwork belongs, provided by the client when initially creating the subnetwork. This field can be set only at resource creation time. "privateIpGoogleAccess": True or False, # Whether the VMs in this subnet can access Google services without assigned external IP addresses. This field can be both set at resource creation time and updated using setPrivateIpGoogleAccess. "privateIpv6GoogleAccess": "A String", # This field is for internal use. This field can be both set at resource creation time and updated using patch. - "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. + "purpose": "A String", # The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. "region": "A String", # URL of the region where the Subnetwork resides. This field can be set only at resource creation time. "reservedInternalRange": "A String", # The URL of the reserved internal range. "role": "A String", # The role of subnetwork. Currently, this field is only used when purpose is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY. The value can be set to ACTIVE or BACKUP. An ACTIVE subnetwork is one that is currently being used for Envoy-based load balancers in a region. A BACKUP subnetwork is one that is ready to be promoted to ACTIVE or is currently draining. This field can be updated with a patch request. diff --git a/docs/dyn/container_v1beta1.projects.locations.clusters.html b/docs/dyn/container_v1beta1.projects.locations.clusters.html index 69cf553abe3..986355d93c6 100644 --- a/docs/dyn/container_v1beta1.projects.locations.clusters.html +++ b/docs/dyn/container_v1beta1.projects.locations.clusters.html @@ -488,6 +488,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -1203,6 +1204,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -1705,6 +1709,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -2420,6 +2425,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -2787,6 +2795,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -3502,6 +3511,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -4721,6 +4733,9 @@

Method Details

"desiredEnableFqdnNetworkPolicy": True or False, # Enable/Disable FQDN Network Policy for the cluster. "desiredEnableMultiNetworking": True or False, # Enable/Disable Multi-Networking for the cluster "desiredEnablePrivateEndpoint": True or False, # Enable/Disable private endpoint for the cluster's master. Deprecated: Use desired_control_plane_endpoints_config.ip_endpoints_config.enable_public_endpoint instead. Note that the value of enable_public_endpoint is reversed: if enable_private_endpoint is false, then enable_public_endpoint will be true. + "desiredEnterpriseConfig": { # DesiredEnterpriseConfig is a wrapper used for updating enterprise_config. # The desired enterprise configuration for the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. + }, "desiredFleet": { # Fleet is the fleet configuration for the cluster. # The desired fleet configuration for the cluster. "membership": "A String", # Output only. The full resource name of the registered fleet membership of the cluster, in the format `//gkehub.googleapis.com/projects/*/locations/*/memberships/*`. "preRegistered": True or False, # Output only. Whether the cluster has been registered through the fleet API. @@ -4863,6 +4878,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "desiredPodAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The desired config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "desiredPodSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # The desired configuration options for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, diff --git a/docs/dyn/container_v1beta1.projects.zones.clusters.html b/docs/dyn/container_v1beta1.projects.zones.clusters.html index bd25a1b0153..bf4166286df 100644 --- a/docs/dyn/container_v1beta1.projects.zones.clusters.html +++ b/docs/dyn/container_v1beta1.projects.zones.clusters.html @@ -586,6 +586,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -1301,6 +1302,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -1803,6 +1807,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -2518,6 +2523,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -2929,6 +2937,7 @@

Method Details

"endpoint": "A String", # Output only. The IP address of this cluster's master endpoint. The endpoint can be accessed from the internet at `https://username:password@endpoint/`. See the `masterAuth` property of this resource for username and password information. "enterpriseConfig": { # EnterpriseConfig is the cluster enterprise configuration. # GKE Enterprise Configuration. "clusterTier": "A String", # Output only. cluster_tier indicates the effective tier of the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. }, "etag": "A String", # This checksum is computed by the server based on the value of cluster fields, and may be sent on update requests to ensure the client has an up-to-date value before proceeding. "expireTime": "A String", # Output only. The time the cluster will be automatically deleted in [RFC3339](https://www.ietf.org/rfc/rfc3339.txt) text format. @@ -3644,6 +3653,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "podAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "podSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # Configuration for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, @@ -4748,6 +4760,9 @@

Method Details

"desiredEnableFqdnNetworkPolicy": True or False, # Enable/Disable FQDN Network Policy for the cluster. "desiredEnableMultiNetworking": True or False, # Enable/Disable Multi-Networking for the cluster "desiredEnablePrivateEndpoint": True or False, # Enable/Disable private endpoint for the cluster's master. Deprecated: Use desired_control_plane_endpoints_config.ip_endpoints_config.enable_public_endpoint instead. Note that the value of enable_public_endpoint is reversed: if enable_private_endpoint is false, then enable_public_endpoint will be true. + "desiredEnterpriseConfig": { # DesiredEnterpriseConfig is a wrapper used for updating enterprise_config. # The desired enterprise configuration for the cluster. + "desiredTier": "A String", # desired_tier specifies the desired tier of the cluster. + }, "desiredFleet": { # Fleet is the fleet configuration for the cluster. # The desired fleet configuration for the cluster. "membership": "A String", # Output only. The full resource name of the registered fleet membership of the cluster, in the format `//gkehub.googleapis.com/projects/*/locations/*/memberships/*`. "preRegistered": True or False, # Output only. Whether the cluster has been registered through the fleet API. @@ -4890,6 +4905,9 @@

Method Details

}, "productName": "A String", # Name of the parent product associated with the cluster. }, + "desiredPodAutoscaling": { # PodAutoscaling is used for configuration of parameters for workload autoscaling. # The desired config for pod autoscaling. + "hpaProfile": "A String", # Selected Horizontal Pod Autoscaling profile. + }, "desiredPodSecurityPolicyConfig": { # Configuration for the PodSecurityPolicy feature. # The desired configuration options for the PodSecurityPolicy feature. "enabled": True or False, # Enable the PodSecurityPolicy controller for this cluster. If enabled, pods must be valid under a PodSecurityPolicy to be created. }, diff --git a/docs/dyn/dataflow_v1b3.projects.html b/docs/dyn/dataflow_v1b3.projects.html index 5d173542642..246d130332c 100644 --- a/docs/dyn/dataflow_v1b3.projects.html +++ b/docs/dyn/dataflow_v1b3.projects.html @@ -243,6 +243,14 @@

Method Details

"totalMs": "A String", # Total active CPU time across all cores (ie., non-idle) in milliseconds since start-up. }, ], + "gpuUsage": [ # Optional. GPU usage samples. + { # Information about the GPU usage on the worker. + "timestamp": "A String", # Required. Timestamp of the measurement. + "utilization": { # Utilization details about the GPU. # Required. Utilization info about the GPU. + "rate": 3.14, # Required. GPU utilization rate of any kernel over the last sample period in the range of [0, 1]. + }, + }, + ], "memoryInfo": [ # Memory utilization samples. { # Information about the memory usage of a worker or a container within a worker. "currentLimitBytes": "A String", # Instantenous memory limit in bytes. diff --git a/docs/dyn/dataflow_v1b3.projects.locations.html b/docs/dyn/dataflow_v1b3.projects.locations.html index ff32d4ce0b6..720ec6a66ff 100644 --- a/docs/dyn/dataflow_v1b3.projects.locations.html +++ b/docs/dyn/dataflow_v1b3.projects.locations.html @@ -221,6 +221,14 @@

Method Details

"totalMs": "A String", # Total active CPU time across all cores (ie., non-idle) in milliseconds since start-up. }, ], + "gpuUsage": [ # Optional. GPU usage samples. + { # Information about the GPU usage on the worker. + "timestamp": "A String", # Required. Timestamp of the measurement. + "utilization": { # Utilization details about the GPU. # Required. Utilization info about the GPU. + "rate": 3.14, # Required. GPU utilization rate of any kernel over the last sample period in the range of [0, 1]. + }, + }, + ], "memoryInfo": [ # Memory utilization samples. { # Information about the memory usage of a worker or a container within a worker. "currentLimitBytes": "A String", # Instantenous memory limit in bytes. diff --git a/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html b/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html index 84327774c72..7a9c5d2e652 100644 --- a/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html +++ b/docs/dyn/datastream_v1.projects.locations.connectionProfiles.html @@ -153,15 +153,30 @@

Method Details

}, "databaseService": "A String", # Required. Database for the Oracle connection. "hostname": "A String", # Required. Hostname for the Oracle connection. + "oracleAsmConfig": { # Configuration for Oracle Automatic Storage Management (ASM) connection. # Optional. Configuration for Oracle ASM connection. + "asmService": "A String", # Required. ASM service name for the Oracle ASM connection. + "connectionAttributes": { # Optional. Connection string attributes + "a_key": "A String", + }, + "hostname": "A String", # Required. Hostname for the Oracle ASM connection. + "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. + }, + "password": "A String", # Required. Password for the Oracle ASM connection. + "port": 42, # Required. Port for the Oracle ASM connection. + "username": "A String", # Required. Username for the Oracle ASM connection. + }, "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. }, "password": "A String", # Optional. Password for the Oracle connection. Mutually exclusive with the `secret_manager_stored_password` field. "port": 42, # Port for the Oracle connection, default value is 1521. + "secretManagerStoredPassword": "A String", # Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field. "username": "A String", # Required. Username for the Oracle connection. }, - "postgresqlProfile": { # PostgreSQL database profile. Next ID: 7. # PostgreSQL Connection Profile configuration. + "postgresqlProfile": { # PostgreSQL database profile. # PostgreSQL Connection Profile configuration. "database": "A String", # Required. Database for the PostgreSQL connection. "hostname": "A String", # Required. Hostname for the PostgreSQL connection. "password": "A String", # Optional. Password for the PostgreSQL connection. Mutually exclusive with the `secret_manager_stored_password` field. @@ -302,15 +317,30 @@

Method Details

}, "databaseService": "A String", # Required. Database for the Oracle connection. "hostname": "A String", # Required. Hostname for the Oracle connection. + "oracleAsmConfig": { # Configuration for Oracle Automatic Storage Management (ASM) connection. # Optional. Configuration for Oracle ASM connection. + "asmService": "A String", # Required. ASM service name for the Oracle ASM connection. + "connectionAttributes": { # Optional. Connection string attributes + "a_key": "A String", + }, + "hostname": "A String", # Required. Hostname for the Oracle ASM connection. + "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. + }, + "password": "A String", # Required. Password for the Oracle ASM connection. + "port": 42, # Required. Port for the Oracle ASM connection. + "username": "A String", # Required. Username for the Oracle ASM connection. + }, "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. }, "password": "A String", # Optional. Password for the Oracle connection. Mutually exclusive with the `secret_manager_stored_password` field. "port": 42, # Port for the Oracle connection, default value is 1521. + "secretManagerStoredPassword": "A String", # Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field. "username": "A String", # Required. Username for the Oracle connection. }, - "postgresqlProfile": { # PostgreSQL database profile. Next ID: 7. # PostgreSQL Connection Profile configuration. + "postgresqlProfile": { # PostgreSQL database profile. # PostgreSQL Connection Profile configuration. "database": "A String", # Required. Database for the PostgreSQL connection. "hostname": "A String", # Required. Hostname for the PostgreSQL connection. "password": "A String", # Optional. Password for the PostgreSQL connection. Mutually exclusive with the `secret_manager_stored_password` field. @@ -598,15 +628,30 @@

Method Details

}, "databaseService": "A String", # Required. Database for the Oracle connection. "hostname": "A String", # Required. Hostname for the Oracle connection. + "oracleAsmConfig": { # Configuration for Oracle Automatic Storage Management (ASM) connection. # Optional. Configuration for Oracle ASM connection. + "asmService": "A String", # Required. ASM service name for the Oracle ASM connection. + "connectionAttributes": { # Optional. Connection string attributes + "a_key": "A String", + }, + "hostname": "A String", # Required. Hostname for the Oracle ASM connection. + "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. + }, + "password": "A String", # Required. Password for the Oracle ASM connection. + "port": 42, # Required. Port for the Oracle ASM connection. + "username": "A String", # Required. Username for the Oracle ASM connection. + }, "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. }, "password": "A String", # Optional. Password for the Oracle connection. Mutually exclusive with the `secret_manager_stored_password` field. "port": 42, # Port for the Oracle connection, default value is 1521. + "secretManagerStoredPassword": "A String", # Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field. "username": "A String", # Required. Username for the Oracle connection. }, - "postgresqlProfile": { # PostgreSQL database profile. Next ID: 7. # PostgreSQL Connection Profile configuration. + "postgresqlProfile": { # PostgreSQL database profile. # PostgreSQL Connection Profile configuration. "database": "A String", # Required. Database for the PostgreSQL connection. "hostname": "A String", # Required. Hostname for the PostgreSQL connection. "password": "A String", # Optional. Password for the PostgreSQL connection. Mutually exclusive with the `secret_manager_stored_password` field. @@ -689,15 +734,30 @@

Method Details

}, "databaseService": "A String", # Required. Database for the Oracle connection. "hostname": "A String", # Required. Hostname for the Oracle connection. + "oracleAsmConfig": { # Configuration for Oracle Automatic Storage Management (ASM) connection. # Optional. Configuration for Oracle ASM connection. + "asmService": "A String", # Required. ASM service name for the Oracle ASM connection. + "connectionAttributes": { # Optional. Connection string attributes + "a_key": "A String", + }, + "hostname": "A String", # Required. Hostname for the Oracle ASM connection. + "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. + }, + "password": "A String", # Required. Password for the Oracle ASM connection. + "port": 42, # Required. Port for the Oracle ASM connection. + "username": "A String", # Required. Username for the Oracle ASM connection. + }, "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. }, "password": "A String", # Optional. Password for the Oracle connection. Mutually exclusive with the `secret_manager_stored_password` field. "port": 42, # Port for the Oracle connection, default value is 1521. + "secretManagerStoredPassword": "A String", # Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field. "username": "A String", # Required. Username for the Oracle connection. }, - "postgresqlProfile": { # PostgreSQL database profile. Next ID: 7. # PostgreSQL Connection Profile configuration. + "postgresqlProfile": { # PostgreSQL database profile. # PostgreSQL Connection Profile configuration. "database": "A String", # Required. Database for the PostgreSQL connection. "hostname": "A String", # Required. Hostname for the PostgreSQL connection. "password": "A String", # Optional. Password for the PostgreSQL connection. Mutually exclusive with the `secret_manager_stored_password` field. @@ -789,15 +849,30 @@

Method Details

}, "databaseService": "A String", # Required. Database for the Oracle connection. "hostname": "A String", # Required. Hostname for the Oracle connection. + "oracleAsmConfig": { # Configuration for Oracle Automatic Storage Management (ASM) connection. # Optional. Configuration for Oracle ASM connection. + "asmService": "A String", # Required. ASM service name for the Oracle ASM connection. + "connectionAttributes": { # Optional. Connection string attributes + "a_key": "A String", + }, + "hostname": "A String", # Required. Hostname for the Oracle ASM connection. + "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. + "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. + "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. + }, + "password": "A String", # Required. Password for the Oracle ASM connection. + "port": 42, # Required. Port for the Oracle ASM connection. + "username": "A String", # Required. Username for the Oracle ASM connection. + }, "oracleSslConfig": { # Oracle SSL configuration information. # Optional. SSL configuration for the Oracle connection. "caCertificate": "A String", # Input only. PEM-encoded certificate of the CA that signed the source database server's certificate. "caCertificateSet": True or False, # Output only. Indicates whether the ca_certificate field has been set for this Connection-Profile. }, "password": "A String", # Optional. Password for the Oracle connection. Mutually exclusive with the `secret_manager_stored_password` field. "port": 42, # Port for the Oracle connection, default value is 1521. + "secretManagerStoredPassword": "A String", # Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field. "username": "A String", # Required. Username for the Oracle connection. }, - "postgresqlProfile": { # PostgreSQL database profile. Next ID: 7. # PostgreSQL Connection Profile configuration. + "postgresqlProfile": { # PostgreSQL database profile. # PostgreSQL Connection Profile configuration. "database": "A String", # Required. Database for the PostgreSQL connection. "hostname": "A String", # Required. Hostname for the PostgreSQL connection. "password": "A String", # Optional. Password for the PostgreSQL connection. Mutually exclusive with the `secret_manager_stored_password` field. diff --git a/docs/dyn/datastream_v1.projects.locations.streams.html b/docs/dyn/datastream_v1.projects.locations.streams.html index 16b169002d4..a02192a5bd5 100644 --- a/docs/dyn/datastream_v1.projects.locations.streams.html +++ b/docs/dyn/datastream_v1.projects.locations.streams.html @@ -273,6 +273,8 @@

Method Details

"name": "A String", # Output only. The stream's name. "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration. + "binaryLogPosition": { # Use Binary log position based replication. # Use Binary log position based replication. + }, "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -298,6 +300,8 @@

Method Details

}, ], }, + "gtid": { # Use GTID based replication. # Use GTID based replication. + }, "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -327,6 +331,14 @@

Method Details

"maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non negative. If not set (or set to 0), the system's default value will be used. }, "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration. + "binaryLogParser": { # Configuration to use Binary Log Parser CDC technique. # Use Binary Log Parser. + "logFileDirectories": { # Configuration to specify the Oracle directories to access the log files. # Use Oracle directories. + "archivedLogDirectory": "A String", # Required. Oracle directory for archived logs. + "onlineLogDirectory": "A String", # Required. Oracle directory for online logs. + }, + "oracleAsmLogFileAccess": { # Configuration to use Oracle ASM to access the log files. # Use Oracle ASM. + }, + }, "dropLargeObjects": { # Configuration to drop large object values. # Drop large object values. }, "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. @@ -379,6 +391,8 @@

Method Details

}, ], }, + "logMiner": { # Configuration to use LogMiner CDC method. # Use LogMiner. + }, "maxConcurrentBackfillTasks": 42, # Maximum number of concurrent backfill tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "streamLargeObjects": { # Configuration to stream large object values. # Stream large object values. @@ -737,6 +751,8 @@

Method Details

"name": "A String", # Output only. The stream's name. "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration. + "binaryLogPosition": { # Use Binary log position based replication. # Use Binary log position based replication. + }, "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -762,6 +778,8 @@

Method Details

}, ], }, + "gtid": { # Use GTID based replication. # Use GTID based replication. + }, "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -791,6 +809,14 @@

Method Details

"maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non negative. If not set (or set to 0), the system's default value will be used. }, "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration. + "binaryLogParser": { # Configuration to use Binary Log Parser CDC technique. # Use Binary Log Parser. + "logFileDirectories": { # Configuration to specify the Oracle directories to access the log files. # Use Oracle directories. + "archivedLogDirectory": "A String", # Required. Oracle directory for archived logs. + "onlineLogDirectory": "A String", # Required. Oracle directory for online logs. + }, + "oracleAsmLogFileAccess": { # Configuration to use Oracle ASM to access the log files. # Use Oracle ASM. + }, + }, "dropLargeObjects": { # Configuration to drop large object values. # Drop large object values. }, "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. @@ -843,6 +869,8 @@

Method Details

}, ], }, + "logMiner": { # Configuration to use LogMiner CDC method. # Use LogMiner. + }, "maxConcurrentBackfillTasks": 42, # Maximum number of concurrent backfill tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "streamLargeObjects": { # Configuration to stream large object values. # Stream large object values. @@ -1140,6 +1168,8 @@

Method Details

"name": "A String", # Output only. The stream's name. "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration. + "binaryLogPosition": { # Use Binary log position based replication. # Use Binary log position based replication. + }, "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -1165,6 +1195,8 @@

Method Details

}, ], }, + "gtid": { # Use GTID based replication. # Use GTID based replication. + }, "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -1194,6 +1226,14 @@

Method Details

"maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non negative. If not set (or set to 0), the system's default value will be used. }, "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration. + "binaryLogParser": { # Configuration to use Binary Log Parser CDC technique. # Use Binary Log Parser. + "logFileDirectories": { # Configuration to specify the Oracle directories to access the log files. # Use Oracle directories. + "archivedLogDirectory": "A String", # Required. Oracle directory for archived logs. + "onlineLogDirectory": "A String", # Required. Oracle directory for online logs. + }, + "oracleAsmLogFileAccess": { # Configuration to use Oracle ASM to access the log files. # Use Oracle ASM. + }, + }, "dropLargeObjects": { # Configuration to drop large object values. # Drop large object values. }, "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. @@ -1246,6 +1286,8 @@

Method Details

}, ], }, + "logMiner": { # Configuration to use LogMiner CDC method. # Use LogMiner. + }, "maxConcurrentBackfillTasks": 42, # Maximum number of concurrent backfill tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "streamLargeObjects": { # Configuration to stream large object values. # Stream large object values. @@ -1550,6 +1592,8 @@

Method Details

"name": "A String", # Output only. The stream's name. "sourceConfig": { # The configuration of the stream source. # Required. Source connection profile configuration. "mysqlSourceConfig": { # MySQL source configuration # MySQL data source configuration. + "binaryLogPosition": { # Use Binary log position based replication. # Use Binary log position based replication. + }, "excludeObjects": { # MySQL database structure # MySQL objects to exclude from the stream. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -1575,6 +1619,8 @@

Method Details

}, ], }, + "gtid": { # Use GTID based replication. # Use GTID based replication. + }, "includeObjects": { # MySQL database structure # MySQL objects to retrieve from the source. "mysqlDatabases": [ # Mysql databases on the server { # MySQL database. @@ -1604,6 +1650,14 @@

Method Details

"maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non negative. If not set (or set to 0), the system's default value will be used. }, "oracleSourceConfig": { # Oracle data source configuration # Oracle data source configuration. + "binaryLogParser": { # Configuration to use Binary Log Parser CDC technique. # Use Binary Log Parser. + "logFileDirectories": { # Configuration to specify the Oracle directories to access the log files. # Use Oracle directories. + "archivedLogDirectory": "A String", # Required. Oracle directory for archived logs. + "onlineLogDirectory": "A String", # Required. Oracle directory for online logs. + }, + "oracleAsmLogFileAccess": { # Configuration to use Oracle ASM to access the log files. # Use Oracle ASM. + }, + }, "dropLargeObjects": { # Configuration to drop large object values. # Drop large object values. }, "excludeObjects": { # Oracle database structure. # Oracle objects to exclude from the stream. @@ -1656,6 +1710,8 @@

Method Details

}, ], }, + "logMiner": { # Configuration to use LogMiner CDC method. # Use LogMiner. + }, "maxConcurrentBackfillTasks": 42, # Maximum number of concurrent backfill tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "maxConcurrentCdcTasks": 42, # Maximum number of concurrent CDC tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used. "streamLargeObjects": { # Configuration to stream large object values. # Stream large object values. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.html b/docs/dyn/dialogflow_v2beta1.projects.locations.html index 4d9ebe608b3..afad2beefdc 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.html @@ -114,6 +114,11 @@

Instance Methods

Returns the operations Resource.

+

+ sipTrunks() +

+

Returns the sipTrunks Resource.

+

statelessSuggestion()

diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.sipTrunks.html b/docs/dyn/dialogflow_v2beta1.projects.locations.sipTrunks.html new file mode 100644 index 00000000000..badb9c690b7 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.sipTrunks.html @@ -0,0 +1,324 @@ + + + +

Dialogflow API . projects . locations . sipTrunks

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates a SipTrunk for a specified location.

+

+ delete(name, x__xgafv=None)

+

Deletes a specified SipTrunk.

+

+ get(name, x__xgafv=None)

+

Retrieves the specified SipTrunk.

+

+ list(parent, pageSize=None, pageToken=None, x__xgafv=None)

+

Returns a list of SipTrunks in the specified location.

+

+ list_next()

+

Retrieves the next page of results.

+

+ patch(name, body=None, updateMask=None, x__xgafv=None)

+

Updates the specified SipTrunk.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates a SipTrunk for a specified location.
+
+Args:
+  parent: string, Required. The location to create a SIP trunk for. Format: `projects//locations/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+  "connections": [ # Output only. Connections of the SIP trunk.
+    { # Represents a connection for SIP Trunk.
+      "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+      "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+        "certificateState": "A String", # Output only. The status of the certificate authentication.
+        "errorMessage": "A String", # The error message provided from SIP trunking auth service
+      },
+      "state": "A String", # Output only. State of the connection.
+      "updateTime": "A String", # Output only. When the connection status changed.
+    },
+  ],
+  "displayName": "A String", # Optional. Human readable alias for this trunk.
+  "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+    "A String",
+  ],
+  "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+  "connections": [ # Output only. Connections of the SIP trunk.
+    { # Represents a connection for SIP Trunk.
+      "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+      "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+        "certificateState": "A String", # Output only. The status of the certificate authentication.
+        "errorMessage": "A String", # The error message provided from SIP trunking auth service
+      },
+      "state": "A String", # Output only. State of the connection.
+      "updateTime": "A String", # Output only. When the connection status changed.
+    },
+  ],
+  "displayName": "A String", # Optional. Human readable alias for this trunk.
+  "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+    "A String",
+  ],
+  "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a specified SipTrunk.
+
+Args:
+  name: string, Required. The name of the SIP trunk to delete. Format: `projects//locations//sipTrunks/`. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); }
+}
+
+ +
+ get(name, x__xgafv=None) +
Retrieves the specified SipTrunk.
+
+Args:
+  name: string, Required. The name of the SIP trunk to delete. Format: `projects//locations//sipTrunks/`. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+  "connections": [ # Output only. Connections of the SIP trunk.
+    { # Represents a connection for SIP Trunk.
+      "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+      "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+        "certificateState": "A String", # Output only. The status of the certificate authentication.
+        "errorMessage": "A String", # The error message provided from SIP trunking auth service
+      },
+      "state": "A String", # Output only. State of the connection.
+      "updateTime": "A String", # Output only. When the connection status changed.
+    },
+  ],
+  "displayName": "A String", # Optional. Human readable alias for this trunk.
+  "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+    "A String",
+  ],
+  "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+}
+
+ +
+ list(parent, pageSize=None, pageToken=None, x__xgafv=None) +
Returns a list of SipTrunks in the specified location.
+
+Args:
+  parent: string, Required. The location to list SIP trunks from. Format: `projects//locations/`. (required)
+  pageSize: integer, Optional. The maximum number of items to return in a single page. By default 100 and at most 1000.
+  pageToken: string, Optional. The next_page_token value returned from a previous list request.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # The response message for SipTrunks.ListSipTrunks.
+  "nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list.
+  "sipTrunks": [ # The list of SIP trunks.
+    { # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+      "connections": [ # Output only. Connections of the SIP trunk.
+        { # Represents a connection for SIP Trunk.
+          "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+          "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+            "certificateState": "A String", # Output only. The status of the certificate authentication.
+            "errorMessage": "A String", # The error message provided from SIP trunking auth service
+          },
+          "state": "A String", # Output only. State of the connection.
+          "updateTime": "A String", # Output only. When the connection status changed.
+        },
+      ],
+      "displayName": "A String", # Optional. Human readable alias for this trunk.
+      "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+        "A String",
+      ],
+      "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+    },
+  ],
+}
+
+ +
+ list_next() +
Retrieves the next page of results.
+
+        Args:
+          previous_request: The request for the previous page. (required)
+          previous_response: The response from the request for the previous page. (required)
+
+        Returns:
+          A request object that you can call 'execute()' on to request the next
+          page. Returns None if there are no more items in the collection.
+        
+
+ +
+ patch(name, body=None, updateMask=None, x__xgafv=None) +
Updates the specified SipTrunk.
+
+Args:
+  name: string, Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+  "connections": [ # Output only. Connections of the SIP trunk.
+    { # Represents a connection for SIP Trunk.
+      "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+      "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+        "certificateState": "A String", # Output only. The status of the certificate authentication.
+        "errorMessage": "A String", # The error message provided from SIP trunking auth service
+      },
+      "state": "A String", # Output only. State of the connection.
+      "updateTime": "A String", # Output only. When the connection status changed.
+    },
+  ],
+  "displayName": "A String", # Optional. Human readable alias for this trunk.
+  "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+    "A String",
+  ],
+  "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+}
+
+  updateMask: string, Optional. The mask to control which fields get updated. If the mask is not present, all fields will be updated.
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.
+  "connections": [ # Output only. Connections of the SIP trunk.
+    { # Represents a connection for SIP Trunk.
+      "connectionId": "A String", # Output only. The unique identifier of the SIP Trunk connection.
+      "errorDetails": { # The error details of Sip Trunk connection authentication. # Output only. The error details for the connection. Only populated when authentication errors occur.
+        "certificateState": "A String", # Output only. The status of the certificate authentication.
+        "errorMessage": "A String", # The error message provided from SIP trunking auth service
+      },
+      "state": "A String", # Output only. State of the connection.
+      "updateTime": "A String", # Output only. When the connection status changed.
+    },
+  ],
+  "displayName": "A String", # Optional. Human readable alias for this trunk.
+  "expectedHostname": [ # Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.
+    "A String",
+  ],
+  "name": "A String", # Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v3.projects.locations.agents.generators.html b/docs/dyn/dialogflow_v3.projects.locations.agents.generators.html index 66b99022658..aca577d85ab 100644 --- a/docs/dyn/dialogflow_v3.projects.locations.agents.generators.html +++ b/docs/dyn/dialogflow_v3.projects.locations.agents.generators.html @@ -112,6 +112,12 @@

Method Details

{ # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -135,6 +141,12 @@

Method Details

{ # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -184,6 +196,12 @@

Method Details

{ # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -218,6 +236,12 @@

Method Details

"generators": [ # The list of generators. There will be a maximum number of items returned based on the page_size field in the request. { # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -259,6 +283,12 @@

Method Details

{ # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -283,6 +313,12 @@

Method Details

{ # Generators contain prompt to be sent to the LLM model to generate text. The prompt can contain parameters which will be resolved before calling the model. It can optionally contain banned phrases to ensure the model responses are safe. "displayName": "A String", # Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment. + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. diff --git a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.generators.html b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.generators.html index 42b1b5864bb..2341bee40a3 100644 --- a/docs/dyn/dialogflow_v3beta1.projects.locations.agents.generators.html +++ b/docs/dyn/dialogflow_v3beta1.projects.locations.agents.generators.html @@ -116,6 +116,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -143,6 +149,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -196,6 +208,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -234,6 +252,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -279,6 +303,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. @@ -307,6 +337,12 @@

Method Details

"model": "A String", # The selected LLM model. "promptText": "A String", # The custom prompt to use. }, + "modelParameter": { # Parameters to be passed to the LLM. If not set, default values will be used. # Parameters passed to the LLM to configure its behavior. + "maxDecodeSteps": 42, # The maximum number of tokens to generate. + "temperature": 3.14, # The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random. + "topK": 42, # If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random. + "topP": 3.14, # If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random. + }, "name": "A String", # The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`. "placeholders": [ # Optional. List of custom placeholders in the prompt text. { # Represents a custom placeholder in the prompt text. diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.html index ef6d9736bc3..03b84517d5b 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.html @@ -79,6 +79,11 @@

Instance Methods

Returns the operations Resource.

+

+ sitemaps() +

+

Returns the sitemaps Resource.

+

targetSites()

diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html new file mode 100644 index 00000000000..f3f67bf53e6 --- /dev/null +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html @@ -0,0 +1,171 @@ + + + +

Discovery Engine API . projects . locations . collections . dataStores . siteSearchEngine . sitemaps

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates a Sitemap.

+

+ delete(name, x__xgafv=None)

+

Deletes a Sitemap.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates a Sitemap.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # A sitemap for the SiteSearchEngine.
+  "createTime": "A String", # Output only. The sitemap's creation time.
+  "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+  "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a Sitemap.
+
+Args:
+  name: string, Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.html index b24d1c601eb..72b9981033e 100644 --- a/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.html +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.html @@ -74,6 +74,11 @@

Discovery Engine API . projects . locations . dataStores . siteSearchEngine

Instance Methods

+

+ sitemaps() +

+

Returns the sitemaps Resource.

+

targetSites()

diff --git a/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.sitemaps.html b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.sitemaps.html new file mode 100644 index 00000000000..37c4af5c892 --- /dev/null +++ b/docs/dyn/discoveryengine_v1alpha.projects.locations.dataStores.siteSearchEngine.sitemaps.html @@ -0,0 +1,202 @@ + + + +

Discovery Engine API . projects . locations . dataStores . siteSearchEngine . sitemaps

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates a Sitemap.

+

+ delete(name, x__xgafv=None)

+

Deletes a Sitemap.

+

+ fetch(parent, matcher_urisMatcher_uris=None, x__xgafv=None)

+

Fetch Sitemaps in a DataStore.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates a Sitemap.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # A sitemap for the SiteSearchEngine.
+  "createTime": "A String", # Output only. The sitemap's creation time.
+  "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+  "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a Sitemap.
+
+Args:
+  name: string, Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ fetch(parent, matcher_urisMatcher_uris=None, x__xgafv=None) +
Fetch Sitemaps in a DataStore.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  matcher_urisMatcher_uris: string, The Sitemap uris. (repeated)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response message for SiteSearchEngineService.FetchSitemaps method.
+  "sitemapsMetadata": [ # List of Sitemaps fetched.
+    { # Contains a Sitemap and its metadata.
+      "sitemap": { # A sitemap for the SiteSearchEngine. # The Sitemap.
+        "createTime": "A String", # Output only. The sitemap's creation time.
+        "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+        "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+      },
+    },
+  ],
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.html b/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.html index aa6ae36cd57..87b403c2866 100644 --- a/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.html +++ b/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.html @@ -79,6 +79,11 @@

Instance Methods

Returns the operations Resource.

+

+ sitemaps() +

+

Returns the sitemaps Resource.

+

targetSites()

diff --git a/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html b/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html new file mode 100644 index 00000000000..c66274c691c --- /dev/null +++ b/docs/dyn/discoveryengine_v1beta.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.html @@ -0,0 +1,171 @@ + + + +

Discovery Engine API . projects . locations . collections . dataStores . siteSearchEngine . sitemaps

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates a Sitemap.

+

+ delete(name, x__xgafv=None)

+

Deletes a Sitemap.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates a Sitemap.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # A sitemap for the SiteSearchEngine.
+  "createTime": "A String", # Output only. The sitemap's creation time.
+  "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+  "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a Sitemap.
+
+Args:
+  name: string, Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.html b/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.html index 580b86f5c43..ef20384a04d 100644 --- a/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.html +++ b/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.html @@ -74,6 +74,11 @@

Discovery Engine API . projects . locations . dataStores . siteSearchEngine

Instance Methods

+

+ sitemaps() +

+

Returns the sitemaps Resource.

+

targetSites()

diff --git a/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.sitemaps.html b/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.sitemaps.html new file mode 100644 index 00000000000..f7479766246 --- /dev/null +++ b/docs/dyn/discoveryengine_v1beta.projects.locations.dataStores.siteSearchEngine.sitemaps.html @@ -0,0 +1,202 @@ + + + +

Discovery Engine API . projects . locations . dataStores . siteSearchEngine . sitemaps

+

Instance Methods

+

+ close()

+

Close httplib2 connections.

+

+ create(parent, body=None, x__xgafv=None)

+

Creates a Sitemap.

+

+ delete(name, x__xgafv=None)

+

Deletes a Sitemap.

+

+ fetch(parent, matcher_urisMatcher_uris=None, x__xgafv=None)

+

Fetch Sitemaps in a DataStore.

+

Method Details

+
+ close() +
Close httplib2 connections.
+
+ +
+ create(parent, body=None, x__xgafv=None) +
Creates a Sitemap.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # A sitemap for the SiteSearchEngine.
+  "createTime": "A String", # Output only. The sitemap's creation time.
+  "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+  "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ delete(name, x__xgafv=None) +
Deletes a Sitemap.
+
+Args:
+  name: string, Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned. (required)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+ +
+ fetch(parent, matcher_urisMatcher_uris=None, x__xgafv=None) +
Fetch Sitemaps in a DataStore.
+
+Args:
+  parent: string, Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`. (required)
+  matcher_urisMatcher_uris: string, The Sitemap uris. (repeated)
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # Response message for SiteSearchEngineService.FetchSitemaps method.
+  "sitemapsMetadata": [ # List of Sitemaps fetched.
+    { # Contains a Sitemap and its metadata.
+      "sitemap": { # A sitemap for the SiteSearchEngine. # The Sitemap.
+        "createTime": "A String", # Output only. The sitemap's creation time.
+        "name": "A String", # Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.
+        "uri": "A String", # Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.
+      },
+    },
+  ],
+}
+
+ + \ No newline at end of file diff --git a/docs/dyn/dlp_v2.infoTypes.html b/docs/dyn/dlp_v2.infoTypes.html index 20eba9c0b4a..dd1bca67c40 100644 --- a/docs/dyn/dlp_v2.infoTypes.html +++ b/docs/dyn/dlp_v2.infoTypes.html @@ -79,7 +79,7 @@

Instance Methods

Close httplib2 connections.

list(filter=None, languageCode=None, locationId=None, parent=None, x__xgafv=None)

-

Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.

+

Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.

Method Details

close() @@ -88,7 +88,7 @@

Method Details

list(filter=None, languageCode=None, locationId=None, parent=None, x__xgafv=None) -
Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.
+  
Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.
 
 Args:
   filter: string, filter to only return infoTypes supported by certain parts of the API. Defaults to supported_by=INSPECT.
diff --git a/docs/dyn/dlp_v2.locations.infoTypes.html b/docs/dyn/dlp_v2.locations.infoTypes.html
index 898e9adf173..93cdc550e63 100644
--- a/docs/dyn/dlp_v2.locations.infoTypes.html
+++ b/docs/dyn/dlp_v2.locations.infoTypes.html
@@ -79,7 +79,7 @@ 

Instance Methods

Close httplib2 connections.

list(parent, filter=None, languageCode=None, locationId=None, x__xgafv=None)

-

Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.

+

Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.

Method Details

close() @@ -88,7 +88,7 @@

Method Details

list(parent, filter=None, languageCode=None, locationId=None, x__xgafv=None) -
Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.
+  
Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.
 
 Args:
   parent: string, The parent resource name. The format of this value is as follows: `locations/{location_id}` (required)
diff --git a/docs/dyn/dlp_v2.organizations.locations.connections.html b/docs/dyn/dlp_v2.organizations.locations.connections.html
index d280ed2c0e4..55fb983bbce 100644
--- a/docs/dyn/dlp_v2.organizations.locations.connections.html
+++ b/docs/dyn/dlp_v2.organizations.locations.connections.html
@@ -117,13 +117,13 @@ 

Method Details

The object takes the form of: { # Request message for CreateConnection. - "connection": { # A data connection to allow DLP to profile data in locations that require additional configuration. # Required. The connection resource. + "connection": { # A data connection to allow the DLP API to profile data in locations that require additional configuration. # Required. The connection resource. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -159,13 +159,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -225,13 +225,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -278,13 +278,13 @@

Method Details

{ # Response message for ListConnections. "connections": [ # List of connections. - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -339,13 +339,13 @@

Method Details

The object takes the form of: { # Request message for UpdateConnection. - "connection": { # A data connection to allow DLP to profile data in locations that require additional configuration. # Required. The connection with new values for the relevant fields. + "connection": { # A data connection to allow the DLP API to profile data in locations that require additional configuration. # Required. The connection with new values for the relevant fields. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -382,13 +382,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -435,13 +435,13 @@

Method Details

{ # Response message for SearchConnections. "connections": [ # List of connections that match the search query. Note that only a subset of the fields will be populated, and only "name" is guaranteed to be set. For full details of a Connection, call GetConnection with the name. - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. diff --git a/docs/dyn/dlp_v2.organizations.locations.discoveryConfigs.html b/docs/dyn/dlp_v2.organizations.locations.discoveryConfigs.html index 02992453534..6428d3b78b9 100644 --- a/docs/dyn/dlp_v2.organizations.locations.discoveryConfigs.html +++ b/docs/dyn/dlp_v2.organizations.locations.discoveryConfigs.html @@ -140,7 +140,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -189,7 +189,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -436,7 +436,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -485,7 +485,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -756,7 +756,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -805,7 +805,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1063,7 +1063,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1112,7 +1112,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1378,7 +1378,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1427,7 +1427,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1675,7 +1675,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1724,7 +1724,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html b/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html index 6d26c5fddaf..bf7c9b87958 100644 --- a/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html +++ b/docs/dyn/dlp_v2.organizations.locations.fileStoreDataProfiles.html @@ -157,7 +157,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -190,7 +190,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -220,7 +220,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -269,7 +269,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -836,7 +836,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -869,7 +869,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -899,7 +899,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -948,7 +948,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html b/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html index c29e389bedc..7be3459848a 100644 --- a/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html +++ b/docs/dyn/dlp_v2.organizations.locations.tableDataProfiles.html @@ -157,7 +157,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -190,7 +190,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -220,7 +220,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -269,7 +269,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -780,7 +780,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -813,7 +813,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -843,7 +843,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -892,7 +892,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/dlp_v2.projects.locations.connections.html b/docs/dyn/dlp_v2.projects.locations.connections.html index bf6d35457d8..54b1332be6e 100644 --- a/docs/dyn/dlp_v2.projects.locations.connections.html +++ b/docs/dyn/dlp_v2.projects.locations.connections.html @@ -117,13 +117,13 @@

Method Details

The object takes the form of: { # Request message for CreateConnection. - "connection": { # A data connection to allow DLP to profile data in locations that require additional configuration. # Required. The connection resource. + "connection": { # A data connection to allow the DLP API to profile data in locations that require additional configuration. # Required. The connection resource. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -159,13 +159,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -225,13 +225,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -278,13 +278,13 @@

Method Details

{ # Response message for ListConnections. "connections": [ # List of connections. - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -339,13 +339,13 @@

Method Details

The object takes the form of: { # Request message for UpdateConnection. - "connection": { # A data connection to allow DLP to profile data in locations that require additional configuration. # Required. The connection with new values for the relevant fields. + "connection": { # A data connection to allow the DLP API to profile data in locations that require additional configuration. # Required. The connection with new values for the relevant fields. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -382,13 +382,13 @@

Method Details

Returns: An object of the form: - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. @@ -435,13 +435,13 @@

Method Details

{ # Response message for SearchConnections. "connections": [ # List of connections that match the search query. Note that only a subset of the fields will be populated, and only "name" is guaranteed to be set. For full details of a Connection, call GetConnection with the name. - { # A data connection to allow DLP to profile data in locations that require additional configuration. + { # A data connection to allow the DLP API to profile data in locations that require additional configuration. "cloudSql": { # Cloud SQL connection properties. # Connect to a Cloud SQL instance. "cloudSqlIam": { # Use IAM authentication to connect. This requires the Cloud SQL IAM feature to be enabled on the instance, which is not the default for Cloud SQL. See https://cloud.google.com/sql/docs/postgres/authentication and https://cloud.google.com/sql/docs/mysql/authentication. # Built-in IAM authentication (must be configured in Cloud SQL). }, "connectionName": "A String", # Optional. Immutable. The Cloud SQL instance for which the connection is defined. Only one connection per instance is allowed. This can only be set at creation time, and cannot be updated. It is an error to use a connection_name from different project or region than the one that holds the connection. For example, a Connection resource for Cloud SQL connection_name `project-id:us-central1:sql-instance` must be created under the parent `projects/project-id/locations/us-central1` "databaseEngine": "A String", # Required. The database engine used by the Cloud SQL instance that this connection configures. - "maxConnections": 42, # Required. DLP will limit its connections to max_connections. Must be 2 or greater. + "maxConnections": 42, # Required. The DLP API will limit its connections to max_connections. Must be 2 or greater. "usernamePassword": { # A credential consisting of a username and password, where the password is stored in a Secret Manager resource. Note: Secret Manager [charges apply](https://cloud.google.com/secret-manager/pricing). # A username and password stored in Secret Manager. "passwordSecretVersionName": "A String", # Required. The name of the Secret Manager resource that stores the password, in the form `projects/project-id/secrets/secret-name/versions/version`. "username": "A String", # Required. The username. diff --git a/docs/dyn/dlp_v2.projects.locations.discoveryConfigs.html b/docs/dyn/dlp_v2.projects.locations.discoveryConfigs.html index 98794497dbb..c960d2ff7ac 100644 --- a/docs/dyn/dlp_v2.projects.locations.discoveryConfigs.html +++ b/docs/dyn/dlp_v2.projects.locations.discoveryConfigs.html @@ -140,7 +140,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -189,7 +189,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -436,7 +436,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -485,7 +485,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -756,7 +756,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -805,7 +805,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1063,7 +1063,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1112,7 +1112,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1378,7 +1378,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1427,7 +1427,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -1675,7 +1675,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -1724,7 +1724,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html b/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html index 3caf137d9a8..0e48a33793c 100644 --- a/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html +++ b/docs/dyn/dlp_v2.projects.locations.fileStoreDataProfiles.html @@ -157,7 +157,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -190,7 +190,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -220,7 +220,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -269,7 +269,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -836,7 +836,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -869,7 +869,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -899,7 +899,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -948,7 +948,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html b/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html index 2d2e8019b1f..ba39d552534 100644 --- a/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html +++ b/docs/dyn/dlp_v2.projects.locations.tableDataProfiles.html @@ -157,7 +157,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -190,7 +190,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -220,7 +220,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -269,7 +269,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. @@ -780,7 +780,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -813,7 +813,7 @@

Method Details

"allAssetInventoryAssets": True or False, # All AWS assets stored in Asset Inventory that didn't match other AWS discovery configs. }, }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "discoveryConfig": { # Configuration for discovery to scan resources for profile generation. Only one discovery configuration may exist per organization, folder, or project. The generated data profiles are retained according to the [data retention policy] (https://cloud.google.com/sensitive-data-protection/docs/data-profiles#retention). # A copy of the configuration used to generate this profile. "actions": [ # Actions to execute at the completion of scanning. @@ -843,7 +843,7 @@

Method Details

}, "publishToChronicle": { # Message expressing intention to publish to Google Security Operations. # Publishes generated data profiles to Google Security Operations. For more information, see [Use Sensitive Data Protection data in context-aware analytics](https://cloud.google.com/chronicle/docs/detection/usecase-dlp-high-risk-user-download). }, - "publishToScc": { # If set, a summary finding will be created/updated in SCC for each profile. # Publishes findings to SCC for each data profile. + "publishToScc": { # If set, a summary finding will be created or updated in Security Command Center for each profile. # Publishes findings to Security Command Center for each data profile. }, "tagResources": { # If set, attaches the [tags] (https://cloud.google.com/resource-manager/docs/tags/tags-overview) provided to profiled resources. Tags support [access control](https://cloud.google.com/iam/docs/tags-access-control). You can conditionally grant or deny access to a resource based on whether the resource has a specific tag. # Tags the profiled resources with the specified tag values. "lowerDataRiskToLow": True or False, # Whether applying a tag to a resource should lower the risk of the profile for that resource. For example, in conjunction with an [IAM deny policy](https://cloud.google.com/iam/docs/deny-overview), you can deny all principals a permission if a tag value is present, mitigating the risk of the resource. This also lowers the data risk of resources at the lower levels of the resource hierarchy. For example, reducing the data risk of a table data profile also reduces the data risk of the constituent column data profiles. @@ -892,7 +892,7 @@

Method Details

"folderId": "A String", # The ID of the folder within an organization to be scanned. "organizationId": "A String", # The ID of an organization to scan. }, - "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled. + "projectId": "A String", # The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled. }, "otherCloudStartingLocation": { # The other cloud starting location for discovery. # Must be set only when scanning other clouds. "awsLocation": { # The AWS starting location for discovery. # The AWS starting location for discovery. diff --git a/docs/dyn/drive_v3.files.accessproposals.html b/docs/dyn/drive_v3.files.accessproposals.html index 9d600ec058d..8941f012b3c 100644 --- a/docs/dyn/drive_v3.files.accessproposals.html +++ b/docs/dyn/drive_v3.files.accessproposals.html @@ -153,7 +153,7 @@

Method Details

action: string, Required. The action to take on the AccessProposal. Allowed values ACTION_UNSPECIFIED - Unspecified action - ACCEPT - The user accepts the proposal + ACCEPT - The user accepts the proposal. Note: If this action is used, the `role` field must have at least one value. DENY - The user denies the proposal role: string, Optional. The roles the approver has allowed, if any. Note: This field is required for the `ACCEPT` action. (repeated) sendNotification: boolean, Optional. Whether to send an email to the requester when the AccessProposal is denied or accepted. diff --git a/docs/dyn/eventarc_v1.projects.locations.pipelines.html b/docs/dyn/eventarc_v1.projects.locations.pipelines.html index 29ddbfd79c9..4053b58bcda 100644 --- a/docs/dyn/eventarc_v1.projects.locations.pipelines.html +++ b/docs/dyn/eventarc_v1.projects.locations.pipelines.html @@ -130,16 +130,16 @@

Method Details

"authenticationConfig": { # Represents a config used to authenticate message requests. # Optional. An authentication config used to authenticate message requests, such that destinations can verify the source. For example, this can be used with private GCP destinations that require GCP credentials to access like Cloud Run. This field is optional and should be set only by users interested in authenticated push "googleOidc": { # Represents a config used to authenticate with a Google OIDC token using a GCP service account. Use this authentication method to invoke your Cloud Run and Cloud Functions destinations or HTTP endpoints that support Google OIDC. # Optional. This authenticate method will apply Google OIDC tokens signed by a GCP service account to the requests. "audience": "A String", # Optional. Audience to be used to generate the OIDC Token. The audience claim identifies the recipient that the JWT is intended for. If unspecified, the destination URI will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. }, "oauthToken": { # Contains information needed for generating an [OAuth token](https://developers.google.com/identity/protocols/OAuth2). This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. # Optional. If specified, an [OAuth token](https://developers.google.com/identity/protocols/OAuth2) will be generated and attached as an `Authorization` header in the HTTP request. This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. "scope": "A String", # Optional. OAuth scope to be used for generating OAuth access token. If not specified, "https://www.googleapis.com/auth/cloud-platform" will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. }, }, "httpEndpoint": { # Represents a HTTP endpoint destination. # Optional. An HTTP endpoint destination described by an URI. If a DNS FQDN is provided as the endpoint, Pipeline will create a peering zone to the consumer VPC and forward DNS requests to the VPC specified by network config to resolve the service endpoint. See: https://cloud.google.com/dns/docs/zones/zones-overview#peering_zones - "messageBindingTemplate": "A String", # Optional. The CEL expression used to construct a new HTTP request to be sent to the final destination. The result of the CEL expression must be a map of key/value pairs such that: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type’s canonical representation. If the `headers` field doesn’t exist then the resulting HTTP request will be without headers. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field doesn’t exist then the resulting HTTP request will be without a body. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. Headers added to the request by previous filters in the chain can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers added by previous mediation steps and not the headers present on the original incoming request. For example, the following CEL expression can be used to construct a Headers-only HTTP request by adding an additional header to the headers added by previous mediations in the Pipeline: ``` {"headers": headers.merge({"new-header-key": "new-header-value"})} ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - toMap: list(map).toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. If a binding is not specified here, by default the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. The Pipeline converts the data field of the message to the format provided in Pipeline.Destination.output_payload_format and maps it to the body field of the result. It also sets the corresponding Content-Type header to the output_payload_format type. If the Pipeline.Destination.output_payload_format is not set, then the Pipeline will treat the data field of the message as opaque binary data and attach it to the request body as bytes. In this case the Content-type header will be set to the value of the datacontenttype attribute set on the incoming CloudEvent message if present and the `application/octet-stream` MIME type otherwise. The Pipeline expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then the outgoing message request may fail with a persistent error. - "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `http://10.10.10.8:80/route`, `http://svc.us-central1.p.local:8080/`. Only HTTP and HTTPS protocols are supported. The host can be either a static IP addressable from the VPC specified by the network config, or an internal DNS hostname of the service resolvable via Cloud DNS. + "messageBindingTemplate": "A String", # Optional. The CEL expression used to modify how the destination-bound HTTP request is constructed. If a binding expression is not specified here, the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. In this representation, all fields except the `data` and `datacontenttype` field on the message are mapped to HTTP request headers with a prefix of `ce-`. To construct the HTTP request payload and the value of the content-type HTTP header, the payload format is defined as follows: 1) Use the output_payload_format_type on the Pipeline.Destination if it is set, else: 2) Use the input_payload_format_type on the Pipeline if it is set, else: 3) Treat the payload as opaque binary data. The `data` field of the message is converted to the payload format or left as-is for case 3) and then attached as the payload of the HTTP request. The `content-type` header on the HTTP request is set to the payload format type or left empty for case 3). However, if a mediation has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header on the HTTP request is set to this `datacontenttype` value. For example, if the `datacontenttype` is "application/json" and the payload format type is "application/json; charset=utf-8", then the `content-type` header on the HTTP request is set to "application/json; charset=utf-8". If a non-empty binding expression is specified then this expression is used to modify the default CloudEvent HTTP Protocol Binding Binary Content representation. The result of the CEL expression must be a map of key/value pairs which is used as follows: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type's canonical representation. If the `headers` field doesn't exist then the resulting HTTP request will be the headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message. Note: If the specified binding expression, has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header in the `headers` map is set to this `datacontenttype` value. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field does not exist then the resulting payload of the HTTP request will be data value of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. Subfields of `message.data` may also be accessed if an input_payload_format has been specified on the Pipeline. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. - Existing headers can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. For example, the following CEL expression can be used to construct an HTTP request by adding an additional header to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message and by overwriting the body of the request: ``` { "headers": headers.merge({"new-header-key": "new-header-value"}), "body": "new-body" } ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. The Pipeline expects that the message it receives adheres to the standard CloudEvent format. If it doesn't then the outgoing message request may fail with a persistent error. + "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `https://svc.us-central1.p.local:8080/route`. Only the HTTPS protocol is supported. }, "messageBus": "A String", # Optional. The resource name of the Message Bus to which events should be published. The Message Bus resource should exist in the same project as the Pipeline. Format: `projects/{project}/locations/{location}/messageBuses/{message_bus}` "networkConfig": { # Represents a network config to be used for destination resolution and connectivity. # Optional. Network config is used to configure how Pipeline resolves and connects to a destination. @@ -180,15 +180,15 @@

Method Details

"mediations": [ # Optional. List of mediation operations to be performed on the message. Currently, only one Transformation operation is allowed in each Pipeline. { # Mediation defines different ways to modify the Pipeline. "transformation": { # Transformation defines the way to transform an incoming message. # Optional. How the Pipeline is to transform messages - "transformationTemplate": "A String", # Optional. The template to apply to transform messages. + "transformationTemplate": "A String", # Optional. The CEL expression template to apply to transform messages. The following CEL extension functions are provided for use in this CEL expression: - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. }, }, ], "name": "A String", # Identifier. The resource name of the Pipeline. Must be unique within the location of the project and must be in `projects/{project}/locations/{location}/pipelines/{pipeline}` format. "retryPolicy": { # The retry policy configuration for the Pipeline. The pipeline exponentially backs off in case the destination is non responsive or returns a retryable error code. The default semantics are as follows: The backoff starts with a 5 second delay and doubles the delay after each failed attempt (10 seconds, 20 seconds, 40 seconds, etc.). The delay is capped at 60 seconds by default. Please note that if you set the min_retry_delay and max_retry_delay fields to the same value this will make the duration between retries constant. # Optional. The retry policy to use in the pipeline. "maxAttempts": 42, # Optional. The maximum number of delivery attempts for any message. The value must be between 1 and 100. The default value for this field is 5. - "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 60. - "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 5. + "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 60. + "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 5. }, "uid": "A String", # Output only. Server-assigned unique identifier for the Pipeline. The value is a UUID4 string and guaranteed to remain unchanged until the resource is deleted. "updateTime": "A String", # Output only. The last-modified time. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z". @@ -288,16 +288,16 @@

Method Details

"authenticationConfig": { # Represents a config used to authenticate message requests. # Optional. An authentication config used to authenticate message requests, such that destinations can verify the source. For example, this can be used with private GCP destinations that require GCP credentials to access like Cloud Run. This field is optional and should be set only by users interested in authenticated push "googleOidc": { # Represents a config used to authenticate with a Google OIDC token using a GCP service account. Use this authentication method to invoke your Cloud Run and Cloud Functions destinations or HTTP endpoints that support Google OIDC. # Optional. This authenticate method will apply Google OIDC tokens signed by a GCP service account to the requests. "audience": "A String", # Optional. Audience to be used to generate the OIDC Token. The audience claim identifies the recipient that the JWT is intended for. If unspecified, the destination URI will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. }, "oauthToken": { # Contains information needed for generating an [OAuth token](https://developers.google.com/identity/protocols/OAuth2). This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. # Optional. If specified, an [OAuth token](https://developers.google.com/identity/protocols/OAuth2) will be generated and attached as an `Authorization` header in the HTTP request. This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. "scope": "A String", # Optional. OAuth scope to be used for generating OAuth access token. If not specified, "https://www.googleapis.com/auth/cloud-platform" will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. }, }, "httpEndpoint": { # Represents a HTTP endpoint destination. # Optional. An HTTP endpoint destination described by an URI. If a DNS FQDN is provided as the endpoint, Pipeline will create a peering zone to the consumer VPC and forward DNS requests to the VPC specified by network config to resolve the service endpoint. See: https://cloud.google.com/dns/docs/zones/zones-overview#peering_zones - "messageBindingTemplate": "A String", # Optional. The CEL expression used to construct a new HTTP request to be sent to the final destination. The result of the CEL expression must be a map of key/value pairs such that: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type’s canonical representation. If the `headers` field doesn’t exist then the resulting HTTP request will be without headers. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field doesn’t exist then the resulting HTTP request will be without a body. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. Headers added to the request by previous filters in the chain can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers added by previous mediation steps and not the headers present on the original incoming request. For example, the following CEL expression can be used to construct a Headers-only HTTP request by adding an additional header to the headers added by previous mediations in the Pipeline: ``` {"headers": headers.merge({"new-header-key": "new-header-value"})} ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - toMap: list(map).toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. If a binding is not specified here, by default the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. The Pipeline converts the data field of the message to the format provided in Pipeline.Destination.output_payload_format and maps it to the body field of the result. It also sets the corresponding Content-Type header to the output_payload_format type. If the Pipeline.Destination.output_payload_format is not set, then the Pipeline will treat the data field of the message as opaque binary data and attach it to the request body as bytes. In this case the Content-type header will be set to the value of the datacontenttype attribute set on the incoming CloudEvent message if present and the `application/octet-stream` MIME type otherwise. The Pipeline expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then the outgoing message request may fail with a persistent error. - "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `http://10.10.10.8:80/route`, `http://svc.us-central1.p.local:8080/`. Only HTTP and HTTPS protocols are supported. The host can be either a static IP addressable from the VPC specified by the network config, or an internal DNS hostname of the service resolvable via Cloud DNS. + "messageBindingTemplate": "A String", # Optional. The CEL expression used to modify how the destination-bound HTTP request is constructed. If a binding expression is not specified here, the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. In this representation, all fields except the `data` and `datacontenttype` field on the message are mapped to HTTP request headers with a prefix of `ce-`. To construct the HTTP request payload and the value of the content-type HTTP header, the payload format is defined as follows: 1) Use the output_payload_format_type on the Pipeline.Destination if it is set, else: 2) Use the input_payload_format_type on the Pipeline if it is set, else: 3) Treat the payload as opaque binary data. The `data` field of the message is converted to the payload format or left as-is for case 3) and then attached as the payload of the HTTP request. The `content-type` header on the HTTP request is set to the payload format type or left empty for case 3). However, if a mediation has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header on the HTTP request is set to this `datacontenttype` value. For example, if the `datacontenttype` is "application/json" and the payload format type is "application/json; charset=utf-8", then the `content-type` header on the HTTP request is set to "application/json; charset=utf-8". If a non-empty binding expression is specified then this expression is used to modify the default CloudEvent HTTP Protocol Binding Binary Content representation. The result of the CEL expression must be a map of key/value pairs which is used as follows: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type's canonical representation. If the `headers` field doesn't exist then the resulting HTTP request will be the headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message. Note: If the specified binding expression, has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header in the `headers` map is set to this `datacontenttype` value. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field does not exist then the resulting payload of the HTTP request will be data value of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. Subfields of `message.data` may also be accessed if an input_payload_format has been specified on the Pipeline. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. - Existing headers can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. For example, the following CEL expression can be used to construct an HTTP request by adding an additional header to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message and by overwriting the body of the request: ``` { "headers": headers.merge({"new-header-key": "new-header-value"}), "body": "new-body" } ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. The Pipeline expects that the message it receives adheres to the standard CloudEvent format. If it doesn't then the outgoing message request may fail with a persistent error. + "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `https://svc.us-central1.p.local:8080/route`. Only the HTTPS protocol is supported. }, "messageBus": "A String", # Optional. The resource name of the Message Bus to which events should be published. The Message Bus resource should exist in the same project as the Pipeline. Format: `projects/{project}/locations/{location}/messageBuses/{message_bus}` "networkConfig": { # Represents a network config to be used for destination resolution and connectivity. # Optional. Network config is used to configure how Pipeline resolves and connects to a destination. @@ -338,15 +338,15 @@

Method Details

"mediations": [ # Optional. List of mediation operations to be performed on the message. Currently, only one Transformation operation is allowed in each Pipeline. { # Mediation defines different ways to modify the Pipeline. "transformation": { # Transformation defines the way to transform an incoming message. # Optional. How the Pipeline is to transform messages - "transformationTemplate": "A String", # Optional. The template to apply to transform messages. + "transformationTemplate": "A String", # Optional. The CEL expression template to apply to transform messages. The following CEL extension functions are provided for use in this CEL expression: - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. }, }, ], "name": "A String", # Identifier. The resource name of the Pipeline. Must be unique within the location of the project and must be in `projects/{project}/locations/{location}/pipelines/{pipeline}` format. "retryPolicy": { # The retry policy configuration for the Pipeline. The pipeline exponentially backs off in case the destination is non responsive or returns a retryable error code. The default semantics are as follows: The backoff starts with a 5 second delay and doubles the delay after each failed attempt (10 seconds, 20 seconds, 40 seconds, etc.). The delay is capped at 60 seconds by default. Please note that if you set the min_retry_delay and max_retry_delay fields to the same value this will make the duration between retries constant. # Optional. The retry policy to use in the pipeline. "maxAttempts": 42, # Optional. The maximum number of delivery attempts for any message. The value must be between 1 and 100. The default value for this field is 5. - "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 60. - "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 5. + "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 60. + "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 5. }, "uid": "A String", # Output only. Server-assigned unique identifier for the Pipeline. The value is a UUID4 string and guaranteed to remain unchanged until the resource is deleted. "updateTime": "A String", # Output only. The last-modified time. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z". @@ -433,16 +433,16 @@

Method Details

"authenticationConfig": { # Represents a config used to authenticate message requests. # Optional. An authentication config used to authenticate message requests, such that destinations can verify the source. For example, this can be used with private GCP destinations that require GCP credentials to access like Cloud Run. This field is optional and should be set only by users interested in authenticated push "googleOidc": { # Represents a config used to authenticate with a Google OIDC token using a GCP service account. Use this authentication method to invoke your Cloud Run and Cloud Functions destinations or HTTP endpoints that support Google OIDC. # Optional. This authenticate method will apply Google OIDC tokens signed by a GCP service account to the requests. "audience": "A String", # Optional. Audience to be used to generate the OIDC Token. The audience claim identifies the recipient that the JWT is intended for. If unspecified, the destination URI will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. }, "oauthToken": { # Contains information needed for generating an [OAuth token](https://developers.google.com/identity/protocols/OAuth2). This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. # Optional. If specified, an [OAuth token](https://developers.google.com/identity/protocols/OAuth2) will be generated and attached as an `Authorization` header in the HTTP request. This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. "scope": "A String", # Optional. OAuth scope to be used for generating OAuth access token. If not specified, "https://www.googleapis.com/auth/cloud-platform" will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. }, }, "httpEndpoint": { # Represents a HTTP endpoint destination. # Optional. An HTTP endpoint destination described by an URI. If a DNS FQDN is provided as the endpoint, Pipeline will create a peering zone to the consumer VPC and forward DNS requests to the VPC specified by network config to resolve the service endpoint. See: https://cloud.google.com/dns/docs/zones/zones-overview#peering_zones - "messageBindingTemplate": "A String", # Optional. The CEL expression used to construct a new HTTP request to be sent to the final destination. The result of the CEL expression must be a map of key/value pairs such that: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type’s canonical representation. If the `headers` field doesn’t exist then the resulting HTTP request will be without headers. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field doesn’t exist then the resulting HTTP request will be without a body. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. Headers added to the request by previous filters in the chain can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers added by previous mediation steps and not the headers present on the original incoming request. For example, the following CEL expression can be used to construct a Headers-only HTTP request by adding an additional header to the headers added by previous mediations in the Pipeline: ``` {"headers": headers.merge({"new-header-key": "new-header-value"})} ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - toMap: list(map).toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. If a binding is not specified here, by default the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. The Pipeline converts the data field of the message to the format provided in Pipeline.Destination.output_payload_format and maps it to the body field of the result. It also sets the corresponding Content-Type header to the output_payload_format type. If the Pipeline.Destination.output_payload_format is not set, then the Pipeline will treat the data field of the message as opaque binary data and attach it to the request body as bytes. In this case the Content-type header will be set to the value of the datacontenttype attribute set on the incoming CloudEvent message if present and the `application/octet-stream` MIME type otherwise. The Pipeline expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then the outgoing message request may fail with a persistent error. - "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `http://10.10.10.8:80/route`, `http://svc.us-central1.p.local:8080/`. Only HTTP and HTTPS protocols are supported. The host can be either a static IP addressable from the VPC specified by the network config, or an internal DNS hostname of the service resolvable via Cloud DNS. + "messageBindingTemplate": "A String", # Optional. The CEL expression used to modify how the destination-bound HTTP request is constructed. If a binding expression is not specified here, the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. In this representation, all fields except the `data` and `datacontenttype` field on the message are mapped to HTTP request headers with a prefix of `ce-`. To construct the HTTP request payload and the value of the content-type HTTP header, the payload format is defined as follows: 1) Use the output_payload_format_type on the Pipeline.Destination if it is set, else: 2) Use the input_payload_format_type on the Pipeline if it is set, else: 3) Treat the payload as opaque binary data. The `data` field of the message is converted to the payload format or left as-is for case 3) and then attached as the payload of the HTTP request. The `content-type` header on the HTTP request is set to the payload format type or left empty for case 3). However, if a mediation has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header on the HTTP request is set to this `datacontenttype` value. For example, if the `datacontenttype` is "application/json" and the payload format type is "application/json; charset=utf-8", then the `content-type` header on the HTTP request is set to "application/json; charset=utf-8". If a non-empty binding expression is specified then this expression is used to modify the default CloudEvent HTTP Protocol Binding Binary Content representation. The result of the CEL expression must be a map of key/value pairs which is used as follows: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type's canonical representation. If the `headers` field doesn't exist then the resulting HTTP request will be the headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message. Note: If the specified binding expression, has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header in the `headers` map is set to this `datacontenttype` value. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field does not exist then the resulting payload of the HTTP request will be data value of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. Subfields of `message.data` may also be accessed if an input_payload_format has been specified on the Pipeline. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. - Existing headers can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. For example, the following CEL expression can be used to construct an HTTP request by adding an additional header to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message and by overwriting the body of the request: ``` { "headers": headers.merge({"new-header-key": "new-header-value"}), "body": "new-body" } ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. The Pipeline expects that the message it receives adheres to the standard CloudEvent format. If it doesn't then the outgoing message request may fail with a persistent error. + "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `https://svc.us-central1.p.local:8080/route`. Only the HTTPS protocol is supported. }, "messageBus": "A String", # Optional. The resource name of the Message Bus to which events should be published. The Message Bus resource should exist in the same project as the Pipeline. Format: `projects/{project}/locations/{location}/messageBuses/{message_bus}` "networkConfig": { # Represents a network config to be used for destination resolution and connectivity. # Optional. Network config is used to configure how Pipeline resolves and connects to a destination. @@ -483,15 +483,15 @@

Method Details

"mediations": [ # Optional. List of mediation operations to be performed on the message. Currently, only one Transformation operation is allowed in each Pipeline. { # Mediation defines different ways to modify the Pipeline. "transformation": { # Transformation defines the way to transform an incoming message. # Optional. How the Pipeline is to transform messages - "transformationTemplate": "A String", # Optional. The template to apply to transform messages. + "transformationTemplate": "A String", # Optional. The CEL expression template to apply to transform messages. The following CEL extension functions are provided for use in this CEL expression: - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. }, }, ], "name": "A String", # Identifier. The resource name of the Pipeline. Must be unique within the location of the project and must be in `projects/{project}/locations/{location}/pipelines/{pipeline}` format. "retryPolicy": { # The retry policy configuration for the Pipeline. The pipeline exponentially backs off in case the destination is non responsive or returns a retryable error code. The default semantics are as follows: The backoff starts with a 5 second delay and doubles the delay after each failed attempt (10 seconds, 20 seconds, 40 seconds, etc.). The delay is capped at 60 seconds by default. Please note that if you set the min_retry_delay and max_retry_delay fields to the same value this will make the duration between retries constant. # Optional. The retry policy to use in the pipeline. "maxAttempts": 42, # Optional. The maximum number of delivery attempts for any message. The value must be between 1 and 100. The default value for this field is 5. - "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 60. - "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 5. + "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 60. + "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 5. }, "uid": "A String", # Output only. Server-assigned unique identifier for the Pipeline. The value is a UUID4 string and guaranteed to remain unchanged until the resource is deleted. "updateTime": "A String", # Output only. The last-modified time. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z". @@ -537,16 +537,16 @@

Method Details

"authenticationConfig": { # Represents a config used to authenticate message requests. # Optional. An authentication config used to authenticate message requests, such that destinations can verify the source. For example, this can be used with private GCP destinations that require GCP credentials to access like Cloud Run. This field is optional and should be set only by users interested in authenticated push "googleOidc": { # Represents a config used to authenticate with a Google OIDC token using a GCP service account. Use this authentication method to invoke your Cloud Run and Cloud Functions destinations or HTTP endpoints that support Google OIDC. # Optional. This authenticate method will apply Google OIDC tokens signed by a GCP service account to the requests. "audience": "A String", # Optional. Audience to be used to generate the OIDC Token. The audience claim identifies the recipient that the JWT is intended for. If unspecified, the destination URI will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests. }, "oauthToken": { # Contains information needed for generating an [OAuth token](https://developers.google.com/identity/protocols/OAuth2). This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. # Optional. If specified, an [OAuth token](https://developers.google.com/identity/protocols/OAuth2) will be generated and attached as an `Authorization` header in the HTTP request. This type of authorization should generally only be used when calling Google APIs hosted on *.googleapis.com. "scope": "A String", # Optional. OAuth scope to be used for generating OAuth access token. If not specified, "https://www.googleapis.com/auth/cloud-platform" will be used. - "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. + "serviceAccount": "A String", # Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests. }, }, "httpEndpoint": { # Represents a HTTP endpoint destination. # Optional. An HTTP endpoint destination described by an URI. If a DNS FQDN is provided as the endpoint, Pipeline will create a peering zone to the consumer VPC and forward DNS requests to the VPC specified by network config to resolve the service endpoint. See: https://cloud.google.com/dns/docs/zones/zones-overview#peering_zones - "messageBindingTemplate": "A String", # Optional. The CEL expression used to construct a new HTTP request to be sent to the final destination. The result of the CEL expression must be a map of key/value pairs such that: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type’s canonical representation. If the `headers` field doesn’t exist then the resulting HTTP request will be without headers. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field doesn’t exist then the resulting HTTP request will be without a body. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. Headers added to the request by previous filters in the chain can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers added by previous mediation steps and not the headers present on the original incoming request. For example, the following CEL expression can be used to construct a Headers-only HTTP request by adding an additional header to the headers added by previous mediations in the Pipeline: ``` {"headers": headers.merge({"new-header-key": "new-header-value"})} ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - toMap: list(map).toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. If a binding is not specified here, by default the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. The Pipeline converts the data field of the message to the format provided in Pipeline.Destination.output_payload_format and maps it to the body field of the result. It also sets the corresponding Content-Type header to the output_payload_format type. If the Pipeline.Destination.output_payload_format is not set, then the Pipeline will treat the data field of the message as opaque binary data and attach it to the request body as bytes. In this case the Content-type header will be set to the value of the datacontenttype attribute set on the incoming CloudEvent message if present and the `application/octet-stream` MIME type otherwise. The Pipeline expects that the content of the message will adhere to the standard CloudEvent format. If it doesn’t then the outgoing message request may fail with a persistent error. - "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `http://10.10.10.8:80/route`, `http://svc.us-central1.p.local:8080/`. Only HTTP and HTTPS protocols are supported. The host can be either a static IP addressable from the VPC specified by the network config, or an internal DNS hostname of the service resolvable via Cloud DNS. + "messageBindingTemplate": "A String", # Optional. The CEL expression used to modify how the destination-bound HTTP request is constructed. If a binding expression is not specified here, the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. In this representation, all fields except the `data` and `datacontenttype` field on the message are mapped to HTTP request headers with a prefix of `ce-`. To construct the HTTP request payload and the value of the content-type HTTP header, the payload format is defined as follows: 1) Use the output_payload_format_type on the Pipeline.Destination if it is set, else: 2) Use the input_payload_format_type on the Pipeline if it is set, else: 3) Treat the payload as opaque binary data. The `data` field of the message is converted to the payload format or left as-is for case 3) and then attached as the payload of the HTTP request. The `content-type` header on the HTTP request is set to the payload format type or left empty for case 3). However, if a mediation has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header on the HTTP request is set to this `datacontenttype` value. For example, if the `datacontenttype` is "application/json" and the payload format type is "application/json; charset=utf-8", then the `content-type` header on the HTTP request is set to "application/json; charset=utf-8". If a non-empty binding expression is specified then this expression is used to modify the default CloudEvent HTTP Protocol Binding Binary Content representation. The result of the CEL expression must be a map of key/value pairs which is used as follows: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type's canonical representation. If the `headers` field doesn't exist then the resulting HTTP request will be the headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message. Note: If the specified binding expression, has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header in the `headers` map is set to this `datacontenttype` value. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field does not exist then the resulting payload of the HTTP request will be data value of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. Subfields of `message.data` may also be accessed if an input_payload_format has been specified on the Pipeline. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. - Existing headers can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. For example, the following CEL expression can be used to construct an HTTP request by adding an additional header to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message and by overwriting the body of the request: ``` { "headers": headers.merge({"new-header-key": "new-header-value"}), "body": "new-body" } ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. The Pipeline expects that the message it receives adheres to the standard CloudEvent format. If it doesn't then the outgoing message request may fail with a persistent error. + "uri": "A String", # Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `https://svc.us-central1.p.local:8080/route`. Only the HTTPS protocol is supported. }, "messageBus": "A String", # Optional. The resource name of the Message Bus to which events should be published. The Message Bus resource should exist in the same project as the Pipeline. Format: `projects/{project}/locations/{location}/messageBuses/{message_bus}` "networkConfig": { # Represents a network config to be used for destination resolution and connectivity. # Optional. Network config is used to configure how Pipeline resolves and connects to a destination. @@ -587,15 +587,15 @@

Method Details

"mediations": [ # Optional. List of mediation operations to be performed on the message. Currently, only one Transformation operation is allowed in each Pipeline. { # Mediation defines different ways to modify the Pipeline. "transformation": { # Transformation defines the way to transform an incoming message. # Optional. How the Pipeline is to transform messages - "transformationTemplate": "A String", # Optional. The template to apply to transform messages. + "transformationTemplate": "A String", # Optional. The CEL expression template to apply to transform messages. The following CEL extension functions are provided for use in this CEL expression: - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are "." separated indices of the map keys. - For example: { "a": 1, "b": { "c": 2, "d": 3 } "e": [4, 5] } .denormalize() -> { "a": 1, "b.c": 2, "b.d": 3, "e.0": 4, "e.1": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. }, }, ], "name": "A String", # Identifier. The resource name of the Pipeline. Must be unique within the location of the project and must be in `projects/{project}/locations/{location}/pipelines/{pipeline}` format. "retryPolicy": { # The retry policy configuration for the Pipeline. The pipeline exponentially backs off in case the destination is non responsive or returns a retryable error code. The default semantics are as follows: The backoff starts with a 5 second delay and doubles the delay after each failed attempt (10 seconds, 20 seconds, 40 seconds, etc.). The delay is capped at 60 seconds by default. Please note that if you set the min_retry_delay and max_retry_delay fields to the same value this will make the duration between retries constant. # Optional. The retry policy to use in the pipeline. "maxAttempts": 42, # Optional. The maximum number of delivery attempts for any message. The value must be between 1 and 100. The default value for this field is 5. - "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 60. - "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 5. + "maxRetryDelay": "A String", # Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 60. + "minRetryDelay": "A String", # Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 5. }, "uid": "A String", # Output only. Server-assigned unique identifier for the Pipeline. The value is a UUID4 string and guaranteed to remain unchanged until the resource is deleted. "updateTime": "A String", # Output only. The last-modified time. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z". diff --git a/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.html b/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.html index 668c2b27cf9..5cc14eb7f45 100644 --- a/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.html +++ b/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.html @@ -109,6 +109,7 @@

Method Details

An object of the form: { # Configuration for automated tests + "displayName": "A String", # Optional. Display name of the AI driven test. Required if the release test is created with multiple goals. "name": "A String", # Identifier. The name of the test configuration resource. Format: `projects/{project_number}/apps/{app_id}/testConfig` "roboCrawler": { # Configuration for Robo crawler # Optional. Configuration for Robo crawler "aiInstructions": { # Instructions for AI driven test # Optional. Instructions for AI driven test @@ -117,6 +118,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -151,6 +154,7 @@

Method Details

The object takes the form of: { # Configuration for automated tests + "displayName": "A String", # Optional. Display name of the AI driven test. Required if the release test is created with multiple goals. "name": "A String", # Identifier. The name of the test configuration resource. Format: `projects/{project_number}/apps/{app_id}/testConfig` "roboCrawler": { # Configuration for Robo crawler # Optional. Configuration for Robo crawler "aiInstructions": { # Instructions for AI driven test # Optional. Instructions for AI driven test @@ -159,6 +163,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -192,6 +198,7 @@

Method Details

An object of the form: { # Configuration for automated tests + "displayName": "A String", # Optional. Display name of the AI driven test. Required if the release test is created with multiple goals. "name": "A String", # Identifier. The name of the test configuration resource. Format: `projects/{project_number}/apps/{app_id}/testConfig` "roboCrawler": { # Configuration for Robo crawler # Optional. Configuration for Robo crawler "aiInstructions": { # Instructions for AI driven test # Optional. Instructions for AI driven test @@ -200,6 +207,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, diff --git a/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.releases.tests.html b/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.releases.tests.html index 87b719a6105..686c3f19b7a 100644 --- a/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.releases.tests.html +++ b/docs/dyn/firebaseappdistribution_v1alpha.projects.apps.releases.tests.html @@ -132,6 +132,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -199,6 +201,8 @@

Method Details

"step": { # A step to be accomplished by the AI # Required. The step performed by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, }, ], @@ -229,7 +233,7 @@

Method Details

"videoUri": "A String", # Output only. A URI to a video of the test run. }, ], - "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals + "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals. "loginCredential": { # Login credential for automated tests # Optional. Input only. Login credentials for the test. Input only. "fieldHints": { # Hints to the crawler for identifying input fields # Optional. Hints to the crawler for identifying input fields "passwordResourceName": "A String", # Required. The Android resource name of the password UI element. For example, in Java: R.string.foo in xml: @string/foo Only the "foo" part is needed. Reference doc: https://developer.android.com/guide/topics/resources/accessing-resources.html @@ -259,6 +263,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -326,6 +332,8 @@

Method Details

"step": { # A step to be accomplished by the AI # Required. The step performed by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, }, ], @@ -356,7 +364,7 @@

Method Details

"videoUri": "A String", # Output only. A URI to a video of the test run. }, ], - "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals + "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals. "loginCredential": { # Login credential for automated tests # Optional. Input only. Login credentials for the test. Input only. "fieldHints": { # Hints to the crawler for identifying input fields # Optional. Hints to the crawler for identifying input fields "passwordResourceName": "A String", # Required. The Android resource name of the password UI element. For example, in Java: R.string.foo in xml: @string/foo Only the "foo" part is needed. Reference doc: https://developer.android.com/guide/topics/resources/accessing-resources.html @@ -392,6 +400,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -459,6 +469,8 @@

Method Details

"step": { # A step to be accomplished by the AI # Required. The step performed by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, }, ], @@ -489,7 +501,7 @@

Method Details

"videoUri": "A String", # Output only. A URI to a video of the test run. }, ], - "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals + "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals. "loginCredential": { # Login credential for automated tests # Optional. Input only. Login credentials for the test. Input only. "fieldHints": { # Hints to the crawler for identifying input fields # Optional. Hints to the crawler for identifying input fields "passwordResourceName": "A String", # Required. The Android resource name of the password UI element. For example, in Java: R.string.foo in xml: @string/foo Only the "foo" part is needed. Reference doc: https://developer.android.com/guide/topics/resources/accessing-resources.html @@ -535,6 +547,8 @@

Method Details

{ # A step to be accomplished by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, ], }, @@ -602,6 +616,8 @@

Method Details

"step": { # A step to be accomplished by the AI # Required. The step performed by the AI "assertion": "A String", # An assertion to be checked by the AI "goal": "A String", # A goal to be accomplished by the AI + "hint": "A String", # Optional. Hint text containing suggestions to help the agent accomplish the goal + "successCriteria": "A String", # Optional. A description of criteria the agent should use to determine if the goal has been successfully completed }, }, ], @@ -632,7 +648,7 @@

Method Details

"videoUri": "A String", # Output only. A URI to a video of the test run. }, ], - "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals + "displayName": "A String", # Optional. Display name of the release test. Required if the release test is created with multiple goals. "loginCredential": { # Login credential for automated tests # Optional. Input only. Login credentials for the test. Input only. "fieldHints": { # Hints to the crawler for identifying input fields # Optional. Hints to the crawler for identifying input fields "passwordResourceName": "A String", # Required. The Android resource name of the password UI element. For example, in Java: R.string.foo in xml: @string/foo Only the "foo" part is needed. Reference doc: https://developer.android.com/guide/topics/resources/accessing-resources.html diff --git a/docs/dyn/gkehub_v1.projects.locations.features.html b/docs/dyn/gkehub_v1.projects.locations.features.html index fe918ac56e4..50d73bc9ba4 100644 --- a/docs/dyn/gkehub_v1.projects.locations.features.html +++ b/docs/dyn/gkehub_v1.projects.locations.features.html @@ -148,6 +148,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -257,6 +258,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -347,6 +349,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -458,6 +461,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -646,6 +650,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1103,6 +1108,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1212,6 +1218,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -1302,6 +1309,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1413,6 +1421,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -1601,6 +1610,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2046,6 +2056,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2155,6 +2166,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -2245,6 +2257,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2356,6 +2369,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -2544,6 +2558,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2944,6 +2959,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3053,6 +3069,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -3143,6 +3160,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3254,6 +3272,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -3442,6 +3461,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkehub_v1alpha.projects.locations.features.html b/docs/dyn/gkehub_v1alpha.projects.locations.features.html index 992f99e21bc..4fe2c139935 100644 --- a/docs/dyn/gkehub_v1alpha.projects.locations.features.html +++ b/docs/dyn/gkehub_v1alpha.projects.locations.features.html @@ -151,6 +151,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -364,6 +365,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -682,6 +684,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1269,6 +1272,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1482,6 +1486,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1800,6 +1805,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2375,6 +2381,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2588,6 +2595,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2906,6 +2914,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3436,6 +3445,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3649,6 +3659,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3967,6 +3978,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkehub_v1beta.projects.locations.features.html b/docs/dyn/gkehub_v1beta.projects.locations.features.html index aa45ba63e5d..c136a9b2ff2 100644 --- a/docs/dyn/gkehub_v1beta.projects.locations.features.html +++ b/docs/dyn/gkehub_v1beta.projects.locations.features.html @@ -151,6 +151,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -260,6 +261,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -362,6 +364,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -473,6 +476,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -670,6 +674,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1142,6 +1147,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1251,6 +1257,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -1353,6 +1360,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1464,6 +1472,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -1661,6 +1670,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2121,6 +2131,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2230,6 +2241,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -2332,6 +2344,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2443,6 +2456,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -2640,6 +2654,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3055,6 +3070,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3164,6 +3180,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -3266,6 +3283,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -3377,6 +3395,7 @@

Method Details

}, }, "mesh": { # **Service Mesh**: Spec for a single Membership for the servicemesh feature # Anthos Service Mesh-specific spec + "configApi": "A String", # Optional. Specifies the API that will be used for configuring the mesh workloads. "controlPlane": "A String", # Deprecated: use `management` instead Enables automatic control plane management. "management": "A String", # Optional. Enables automatic Service Mesh management. }, @@ -3574,6 +3593,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkehub_v2.projects.locations.memberships.features.html b/docs/dyn/gkehub_v2.projects.locations.memberships.features.html index 342416fbeeb..2489cd334d7 100644 --- a/docs/dyn/gkehub_v2.projects.locations.memberships.features.html +++ b/docs/dyn/gkehub_v2.projects.locations.memberships.features.html @@ -158,6 +158,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -467,6 +468,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -820,6 +822,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1129,6 +1132,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1422,6 +1426,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1731,6 +1736,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2033,6 +2039,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2342,6 +2349,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkehub_v2alpha.projects.locations.memberships.features.html b/docs/dyn/gkehub_v2alpha.projects.locations.memberships.features.html index e22bb4c5108..d0d6fb941de 100644 --- a/docs/dyn/gkehub_v2alpha.projects.locations.memberships.features.html +++ b/docs/dyn/gkehub_v2alpha.projects.locations.memberships.features.html @@ -158,6 +158,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -467,6 +468,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -820,6 +822,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1129,6 +1132,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1422,6 +1426,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1731,6 +1736,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2033,6 +2039,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2342,6 +2349,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkehub_v2beta.projects.locations.memberships.features.html b/docs/dyn/gkehub_v2beta.projects.locations.memberships.features.html index d79621904cb..b0e335e43dd 100644 --- a/docs/dyn/gkehub_v2beta.projects.locations.memberships.features.html +++ b/docs/dyn/gkehub_v2beta.projects.locations.memberships.features.html @@ -158,6 +158,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -467,6 +468,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -820,6 +822,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1129,6 +1132,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1422,6 +1426,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -1731,6 +1736,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2033,6 +2039,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. @@ -2342,6 +2349,7 @@

Method Details

}, "preventDrift": True or False, # Set to true to enable the Config Sync admission webhook to prevent drifts. If set to `false`, disables the Config Sync admission webhook and does not prevent drifts. "sourceFormat": "A String", # Specifies whether the Config Sync Repo is in "hierarchical" or "unstructured" mode. + "stopSyncing": True or False, # Set to true to stop syncing configs for a single cluster. Default to false. }, "hierarchyController": { # Configuration for Hierarchy Controller. # Hierarchy Controller configuration for the cluster. Deprecated: Configuring Hierarchy Controller through the configmanagement feature is no longer recommended. Use https://github.com/kubernetes-sigs/hierarchical-namespaces instead. "enableHierarchicalResourceQuota": True or False, # Whether hierarchical resource quota is enabled in this cluster. diff --git a/docs/dyn/gkeonprem_v1.projects.locations.vmwareAdminClusters.html b/docs/dyn/gkeonprem_v1.projects.locations.vmwareAdminClusters.html index c8d4dc6d152..87587ef12e1 100644 --- a/docs/dyn/gkeonprem_v1.projects.locations.vmwareAdminClusters.html +++ b/docs/dyn/gkeonprem_v1.projects.locations.vmwareAdminClusters.html @@ -82,6 +82,9 @@

Instance Methods

close()

Close httplib2 connections.

+

+ create(parent, body, allowPreflightFailure=None, validateOnly=None, vmwareAdminClusterId=None, x__xgafv=None)

+

Creates a new VMware admin cluster in a given project and location. The API needs to be combined with creating a bootstrap cluster to work.

enroll(parent, body=None, x__xgafv=None)

Enrolls an existing VMware admin cluster to the Anthos On-Prem API within a given project and location. Through enrollment, an existing admin cluster will become Anthos On-Prem API managed. The corresponding GCP resources will be created and all future modifications to the cluster will be expected to be performed through the API.

@@ -115,6 +118,283 @@

Method Details

Close httplib2 connections.
+
+ create(parent, body, allowPreflightFailure=None, validateOnly=None, vmwareAdminClusterId=None, x__xgafv=None) +
Creates a new VMware admin cluster in a given project and location. The API needs to be combined with creating a bootstrap cluster to work.
+
+Args:
+  parent: string, Required. The parent of the project and location where the cluster is created in. Format: "projects/{project}/locations/{location}" (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Resource that represents a VMware admin cluster.
+  "addonNode": { # VmwareAdminAddonNodeConfig contains add-on node configurations for VMware admin cluster. # The VMware admin cluster addon node configuration.
+    "autoResizeConfig": { # Represents auto resizing configurations for the VMware user cluster. # VmwareAutoResizeConfig config specifies auto resize config.
+      "enabled": True or False, # Whether to enable controle plane node auto resizing.
+    },
+  },
+  "annotations": { # Annotations on the VMware admin cluster. This field has the same restrictions as Kubernetes annotations. The total size of all keys and values combined is limited to 256k. Key can have 2 segments: prefix (optional) and name (required), separated by a slash (/). Prefix must be a DNS subdomain. Name must be 63 characters or less, begin and end with alphanumerics, with dashes (-), underscores (_), dots (.), and alphanumerics between.
+    "a_key": "A String",
+  },
+  "antiAffinityGroups": { # Specifies anti affinity group config for the VMware user cluster. # The VMware admin cluster anti affinity group configuration.
+    "aagConfigDisabled": True or False, # Spread nodes across at least three physical hosts (requires at least three hosts). Enabled by default.
+  },
+  "authorization": { # VmwareAdminAuthorizationConfig represents configuration for admin cluster authorization. # The VMware admin cluster authorization configuration.
+    "viewerUsers": [ # For VMware admin clusters, users will be granted the cluster-viewer role on the cluster.
+      { # ClusterUser configures user principals for an RBAC policy.
+        "username": "A String", # Required. The name of the user, e.g. `my-gcp-id@gmail.com`.
+      },
+    ],
+  },
+  "autoRepairConfig": { # Specifies config to enable/disable auto repair. The cluster-health-controller is deployed only if Enabled is true. # The VMware admin cluster auto repair configuration.
+    "enabled": True or False, # Whether auto repair is enabled.
+  },
+  "bootstrapClusterMembership": "A String", # The bootstrap cluster this VMware admin cluster belongs to.
+  "controlPlaneNode": { # VmwareAdminControlPlaneNodeConfig contains control plane node configuration for VMware admin cluster. # The VMware admin cluster control plane node configuration.
+    "cpus": "A String", # The number of vCPUs for the control-plane node of the admin cluster.
+    "memory": "A String", # The number of mebibytes of memory for the control-plane node of the admin cluster.
+    "replicas": "A String", # The number of control plane nodes for this VMware admin cluster. (default: 1 replica).
+  },
+  "createTime": "A String", # Output only. The time at which VMware admin cluster was created.
+  "description": "A String", # A human readable description of this VMware admin cluster.
+  "endpoint": "A String", # Output only. The DNS name of VMware admin cluster's API server.
+  "etag": "A String", # This checksum is computed by the server based on the value of other fields, and may be sent on update and delete requests to ensure the client has an up-to-date value before proceeding. Allows clients to perform consistent read-modify-writes through optimistic concurrency control.
+  "fleet": { # Fleet related configuration. Fleets are a Google Cloud concept for logically organizing clusters, letting you use and manage multi-cluster capabilities and apply consistent policies across your systems. See [Anthos Fleets](`https://cloud.google.com/anthos/multicluster-management/fleets`) for more details on Anthos multi-cluster capabilities using Fleets. ## # Output only. Fleet configuration for the cluster.
+    "membership": "A String", # Output only. The name of the managed fleet Membership resource associated to this cluster. Membership names are formatted as `projects//locations//memberships/`.
+  },
+  "imageType": "A String", # The OS image type for the VMware admin cluster.
+  "loadBalancer": { # VmwareAdminLoadBalancerConfig contains load balancer configuration for VMware admin cluster. # The VMware admin cluster load balancer configuration.
+    "f5Config": { # VmwareAdminF5BigIpConfig represents configuration parameters for an F5 BIG-IP load balancer. # Configuration for F5 Big IP typed load balancers.
+      "address": "A String", # The load balancer's IP address.
+      "partition": "A String", # The preexisting partition to be used by the load balancer. This partition is usually created for the admin cluster for example: 'my-f5-admin-partition'.
+      "snatPool": "A String", # The pool name. Only necessary, if using SNAT.
+    },
+    "manualLbConfig": { # Manually configured load balancers.
+      "addonsNodePort": 42, # NodePort for add-ons server in the admin cluster.
+      "controlPlaneNodePort": 42, # NodePort for control plane service. The Kubernetes API server in the admin cluster is implemented as a Service of type NodePort (ex. 30968).
+      "ingressHttpNodePort": 42, # NodePort for ingress service's http. The ingress service in the admin cluster is implemented as a Service of type NodePort (ex. 32527).
+      "ingressHttpsNodePort": 42, # NodePort for ingress service's https. The ingress service in the admin cluster is implemented as a Service of type NodePort (ex. 30139).
+      "konnectivityServerNodePort": 42, # NodePort for konnectivity server service running as a sidecar in each kube-apiserver pod (ex. 30564).
+    },
+    "metalLbConfig": { # VmwareAdminMetalLbConfig represents configuration parameters for a MetalLB load balancer. For admin clusters, currently no configurations is needed. # MetalLB load balancers.
+    },
+    "seesawConfig": { # VmwareSeesawConfig represents configuration parameters for an already existing Seesaw load balancer. IMPORTANT: Please note that the Anthos On-Prem API will not generate or update Seesaw configurations it can only bind a pre-existing configuration to a new user cluster. IMPORTANT: When attempting to create a user cluster with a pre-existing Seesaw load balancer you will need to follow some preparation steps before calling the 'CreateVmwareCluster' API method. First you will need to create the user cluster's namespace via kubectl. The namespace will need to use the following naming convention : -gke-onprem-mgmt or -gke-onprem-mgmt depending on whether you used the 'VmwareCluster.local_name' to disambiguate collisions; for more context see the documentation of 'VmwareCluster.local_name'. Once the namespace is created you will need to create a secret resource via kubectl. This secret will contain copies of your Seesaw credentials. The Secret must be called 'user-cluster-creds' and contain Seesaw's SSH and Cert credentials. The credentials must be keyed with the following names: 'seesaw-ssh-private-key', 'seesaw-ssh-public-key', 'seesaw-ssh-ca-key', 'seesaw-ssh-ca-cert'. # Output only. Configuration for Seesaw typed load balancers.
+      "enableHa": True or False, # Enable two load balancer VMs to achieve a highly-available Seesaw load balancer.
+      "group": "A String", # In general the following format should be used for the Seesaw group name: seesaw-for-[cluster_name].
+      "ipBlocks": [ # The IP Blocks to be used by the Seesaw load balancer
+        { # Represents a collection of IP addresses to assign to nodes.
+          "gateway": "A String", # The network gateway used by the VMware user cluster.
+          "ips": [ # The node's network configurations used by the VMware user cluster.
+            { # Represents VMware user cluster node's network configuration.
+              "hostname": "A String", # Hostname of the machine. VM's name will be used if this field is empty.
+              "ip": "A String", # IP could be an IP address (like 1.2.3.4) or a CIDR (like 1.2.3.0/24).
+            },
+          ],
+          "netmask": "A String", # The netmask used by the VMware user cluster.
+        },
+      ],
+      "masterIp": "A String", # MasterIP is the IP announced by the master of Seesaw group.
+      "stackdriverName": "A String", # Name to be used by Stackdriver.
+      "vms": [ # Names of the VMs created for this Seesaw group.
+        "A String",
+      ],
+    },
+    "vipConfig": { # VmwareAdminVipConfig for VMware load balancer configurations. # The VIPs used by the load balancer.
+      "addonsVip": "A String", # The VIP to configure the load balancer for add-ons.
+      "controlPlaneVip": "A String", # The VIP which you previously set aside for the Kubernetes API of the admin cluster.
+    },
+  },
+  "localName": "A String", # Output only. The object name of the VMware OnPremAdminCluster custom resource. This field is used to support conflicting names when enrolling existing clusters to the API. When used as a part of cluster enrollment, this field will differ from the ID in the resource name. For new clusters, this field will match the user provided cluster name and be visible in the last component of the resource name. It is not modifiable. All users should use this name to access their cluster using gkectl or kubectl and should expect to see the local name when viewing admin cluster controller logs.
+  "name": "A String", # Immutable. The VMware admin cluster resource name.
+  "networkConfig": { # VmwareAdminNetworkConfig contains network configuration for VMware admin cluster. # The VMware admin cluster network configuration.
+    "dhcpIpConfig": { # Represents the network configuration required for the VMware user clusters with DHCP IP configurations. # Configuration settings for a DHCP IP configuration.
+      "enabled": True or False, # enabled is a flag to mark if DHCP IP allocation is used for VMware user clusters.
+    },
+    "haControlPlaneConfig": { # Specifies HA admin control plane config. # Configuration for HA admin cluster control plane.
+      "controlPlaneIpBlock": { # Represents a collection of IP addresses to assign to nodes. # Static IP addresses for the admin control plane nodes.
+        "gateway": "A String", # The network gateway used by the VMware user cluster.
+        "ips": [ # The node's network configurations used by the VMware user cluster.
+          { # Represents VMware user cluster node's network configuration.
+            "hostname": "A String", # Hostname of the machine. VM's name will be used if this field is empty.
+            "ip": "A String", # IP could be an IP address (like 1.2.3.4) or a CIDR (like 1.2.3.0/24).
+          },
+        ],
+        "netmask": "A String", # The netmask used by the VMware user cluster.
+      },
+    },
+    "hostConfig": { # Represents the common parameters for all the hosts irrespective of their IP address. # Represents common network settings irrespective of the host's IP address.
+      "dnsSearchDomains": [ # DNS search domains.
+        "A String",
+      ],
+      "dnsServers": [ # DNS servers.
+        "A String",
+      ],
+      "ntpServers": [ # NTP servers.
+        "A String",
+      ],
+    },
+    "podAddressCidrBlocks": [ # Required. All pods in the cluster are assigned an RFC1918 IPv4 address from these ranges. Only a single range is supported. This field cannot be changed after creation.
+      "A String",
+    ],
+    "serviceAddressCidrBlocks": [ # Required. All services in the cluster are assigned an RFC1918 IPv4 address from these ranges. Only a single range is supported. This field cannot be changed after creation.
+      "A String",
+    ],
+    "staticIpConfig": { # Represents the network configuration required for the VMware user clusters with Static IP configurations. # Configuration settings for a static IP configuration.
+      "ipBlocks": [ # Represents the configuration values for static IP allocation to nodes.
+        { # Represents a collection of IP addresses to assign to nodes.
+          "gateway": "A String", # The network gateway used by the VMware user cluster.
+          "ips": [ # The node's network configurations used by the VMware user cluster.
+            { # Represents VMware user cluster node's network configuration.
+              "hostname": "A String", # Hostname of the machine. VM's name will be used if this field is empty.
+              "ip": "A String", # IP could be an IP address (like 1.2.3.4) or a CIDR (like 1.2.3.0/24).
+            },
+          ],
+          "netmask": "A String", # The netmask used by the VMware user cluster.
+        },
+      ],
+    },
+    "vcenterNetwork": "A String", # vcenter_network specifies vCenter network name.
+  },
+  "onPremVersion": "A String", # The Anthos clusters on the VMware version for the admin cluster.
+  "platformConfig": { # VmwarePlatformConfig represents configuration for the VMware platform. # The VMware platform configuration.
+    "bundles": [ # Output only. The list of bundles installed in the admin cluster.
+      { # VmwareBundleConfig represents configuration for the bundle.
+        "status": { # ResourceStatus describes why a cluster or node pool has a certain status. (e.g., ERROR or DEGRADED). # Output only. Resource status for the bundle.
+          "conditions": [ # ResourceCondition provide a standard mechanism for higher-level status reporting from controller.
+            { # ResourceCondition provides a standard mechanism for higher-level status reporting from controller.
+              "lastTransitionTime": "A String", # Last time the condition transit from one status to another.
+              "message": "A String", # Human-readable message indicating details about last transition.
+              "reason": "A String", # Machine-readable message indicating details about last transition.
+              "state": "A String", # state of the condition.
+              "type": "A String", # Type of the condition. (e.g., ClusterRunning, NodePoolRunning or ServerSidePreflightReady)
+            },
+          ],
+          "errorMessage": "A String", # Human-friendly representation of the error message from controller. The error message can be temporary as the controller controller creates a cluster or node pool. If the error message persists for a longer period of time, it can be used to surface error message to indicate real problems requiring user intervention.
+          "version": "A String", # Reflect current version of the resource.
+          "versions": { # Versions describes the mapping of a given version to the number of machines under this version. # Shows the mapping of a given version to the number of machines under this version.
+            "versions": [ # Shows the mapping of a given version to the number of machines under this version.
+              { # Version describes the number of nodes at a given version under a resource.
+                "count": "A String", # Number of machines under the above version.
+                "version": "A String", # Resource version.
+              },
+            ],
+          },
+        },
+        "version": "A String", # The version of the bundle.
+      },
+    ],
+    "platformVersion": "A String", # Output only. The platform version e.g. 1.13.2.
+    "requiredPlatformVersion": "A String", # Input only. The required platform version e.g. 1.13.1. If the current platform version is lower than the target version, the platform version will be updated to the target version. If the target version is not installed in the platform (bundle versions), download the target version bundle.
+    "status": { # ResourceStatus describes why a cluster or node pool has a certain status. (e.g., ERROR or DEGRADED). # Output only. Resource status for the platform.
+      "conditions": [ # ResourceCondition provide a standard mechanism for higher-level status reporting from controller.
+        { # ResourceCondition provides a standard mechanism for higher-level status reporting from controller.
+          "lastTransitionTime": "A String", # Last time the condition transit from one status to another.
+          "message": "A String", # Human-readable message indicating details about last transition.
+          "reason": "A String", # Machine-readable message indicating details about last transition.
+          "state": "A String", # state of the condition.
+          "type": "A String", # Type of the condition. (e.g., ClusterRunning, NodePoolRunning or ServerSidePreflightReady)
+        },
+      ],
+      "errorMessage": "A String", # Human-friendly representation of the error message from controller. The error message can be temporary as the controller controller creates a cluster or node pool. If the error message persists for a longer period of time, it can be used to surface error message to indicate real problems requiring user intervention.
+      "version": "A String", # Reflect current version of the resource.
+      "versions": { # Versions describes the mapping of a given version to the number of machines under this version. # Shows the mapping of a given version to the number of machines under this version.
+        "versions": [ # Shows the mapping of a given version to the number of machines under this version.
+          { # Version describes the number of nodes at a given version under a resource.
+            "count": "A String", # Number of machines under the above version.
+            "version": "A String", # Resource version.
+          },
+        ],
+      },
+    },
+  },
+  "preparedSecrets": { # VmwareAdminPreparedSecretsConfig represents configuration for admin cluster prepared secrets. # Output only. The VMware admin cluster prepared secrets configuration. It should always be enabled by the Central API, instead of letting users set it.
+    "enabled": True or False, # Whether prepared secrets is enabled.
+  },
+  "reconciling": True or False, # Output only. If set, there are currently changes in flight to the VMware admin cluster.
+  "state": "A String", # Output only. The current state of VMware admin cluster.
+  "status": { # ResourceStatus describes why a cluster or node pool has a certain status. (e.g., ERROR or DEGRADED). # Output only. ResourceStatus representing detailed cluster state.
+    "conditions": [ # ResourceCondition provide a standard mechanism for higher-level status reporting from controller.
+      { # ResourceCondition provides a standard mechanism for higher-level status reporting from controller.
+        "lastTransitionTime": "A String", # Last time the condition transit from one status to another.
+        "message": "A String", # Human-readable message indicating details about last transition.
+        "reason": "A String", # Machine-readable message indicating details about last transition.
+        "state": "A String", # state of the condition.
+        "type": "A String", # Type of the condition. (e.g., ClusterRunning, NodePoolRunning or ServerSidePreflightReady)
+      },
+    ],
+    "errorMessage": "A String", # Human-friendly representation of the error message from controller. The error message can be temporary as the controller controller creates a cluster or node pool. If the error message persists for a longer period of time, it can be used to surface error message to indicate real problems requiring user intervention.
+    "version": "A String", # Reflect current version of the resource.
+    "versions": { # Versions describes the mapping of a given version to the number of machines under this version. # Shows the mapping of a given version to the number of machines under this version.
+      "versions": [ # Shows the mapping of a given version to the number of machines under this version.
+        { # Version describes the number of nodes at a given version under a resource.
+          "count": "A String", # Number of machines under the above version.
+          "version": "A String", # Resource version.
+        },
+      ],
+    },
+  },
+  "uid": "A String", # Output only. The unique identifier of the VMware admin cluster.
+  "updateTime": "A String", # Output only. The time at which VMware admin cluster was last updated.
+  "validationCheck": { # ValidationCheck represents the result of preflight check. # Output only. ValidationCheck represents the result of the preflight check job.
+    "option": "A String", # Options used for the validation check
+    "scenario": "A String", # Output only. The scenario when the preflight checks were run.
+    "status": { # ValidationCheckStatus defines the detailed validation check status. # Output only. The detailed validation check status.
+      "result": [ # Individual checks which failed as part of the Preflight check execution.
+        { # ValidationCheckResult defines the details about the validation check.
+          "category": "A String", # The category of the validation.
+          "description": "A String", # The description of the validation check.
+          "details": "A String", # Detailed failure information, which might be unformatted.
+          "reason": "A String", # A human-readable message of the check failure.
+          "state": "A String", # The validation check state.
+        },
+      ],
+    },
+  },
+  "vcenter": { # VmwareAdminVCenterConfig contains VCenter configuration for VMware admin cluster. # The VMware admin cluster VCenter configuration.
+    "address": "A String", # The vCenter IP address.
+    "caCertData": "A String", # Contains the vCenter CA certificate public key for SSL verification.
+    "cluster": "A String", # The name of the vCenter cluster for the admin cluster.
+    "dataDisk": "A String", # The name of the virtual machine disk (VMDK) for the admin cluster.
+    "datacenter": "A String", # The name of the vCenter datacenter for the admin cluster.
+    "datastore": "A String", # The name of the vCenter datastore for the admin cluster.
+    "folder": "A String", # The name of the vCenter folder for the admin cluster.
+    "resourcePool": "A String", # The name of the vCenter resource pool for the admin cluster.
+    "storagePolicyName": "A String", # The name of the vCenter storage policy for the user cluster.
+  },
+}
+
+  allowPreflightFailure: boolean, Optional. If set to true, CLM will force CCFE to persist the cluster resource in RMS when the creation fails during standalone preflight checks. In that case the subsequent create call will fail with "cluster already exists" error and hence a update cluster is required to fix the cluster.
+  validateOnly: boolean, Validate the request without actually doing any updates.
+  vmwareAdminClusterId: string, Required. User provided identifier that is used as part of the resource name; must conform to RFC-1034 and additionally restrict to lower-cased letters. This comes out roughly to: /^a-z+[a-z0-9]$/
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
enroll(parent, body=None, x__xgafv=None)
Enrolls an existing VMware admin cluster to the Anthos On-Prem API within a given project and location. Through enrollment, an existing admin cluster will become Anthos On-Prem API managed. The corresponding GCP resources will be created and all future modifications to the cluster will be expected to be performed through the API.
diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.businessInfo.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.businessInfo.html
index cfe8c50a5d5..beb1d052c57 100644
--- a/docs/dyn/merchantapi_accounts_v1beta.accounts.businessInfo.html
+++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.businessInfo.html
@@ -103,8 +103,8 @@ 

Method Details

Returns: An object of the form: - { # Collection of information related to a business. - "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. + { # The `BusinessInfo` message contains essential information about a merchant's business. This message captures key business details such as physical address, customer service contacts, and region-specific identifiers. + "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. Only `region_code`, `address_lines`, `postal_code`, `administrative_area` and `locality` fields are supported. All other fields are ignored. "addressLines": [ # Unstructured address lines describing the lower levels of an address. Because values in address_lines do not have type information and may sometimes contain multiple values in a single field (e.g. "Austin, TX"), it is important that the line order is clear. The order of address lines should be "envelope order" for the country/region of the address. In places where this can vary (e.g. Japan), address_language is used to make it explicit (e.g. "ja" for large-to-small ordering and "ja-Latn" or "en" for small-to-large). This way, the most specific line of an address can be selected based on the language. The minimum permitted structural representation of an address consists of a region_code with all remaining information placed in the address_lines. It would be possible to format such an address very approximately without geocoding, but no semantic reasoning could be made about any of the address components until it was at least partially resolved. Creating an address only containing a region_code and address_lines, and then geocoding is the recommended way to handle completely unstructured addresses (as opposed to guessing which parts of the address should be localities or administrative areas). "A String", ], @@ -156,8 +156,8 @@

Method Details

body: object, The request body. The object takes the form of: -{ # Collection of information related to a business. - "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. +{ # The `BusinessInfo` message contains essential information about a merchant's business. This message captures key business details such as physical address, customer service contacts, and region-specific identifiers. + "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. Only `region_code`, `address_lines`, `postal_code`, `administrative_area` and `locality` fields are supported. All other fields are ignored. "addressLines": [ # Unstructured address lines describing the lower levels of an address. Because values in address_lines do not have type information and may sometimes contain multiple values in a single field (e.g. "Austin, TX"), it is important that the line order is clear. The order of address lines should be "envelope order" for the country/region of the address. In places where this can vary (e.g. Japan), address_language is used to make it explicit (e.g. "ja" for large-to-small ordering and "ja-Latn" or "en" for small-to-large). This way, the most specific line of an address can be selected based on the language. The minimum permitted structural representation of an address consists of a region_code with all remaining information placed in the address_lines. It would be possible to format such an address very approximately without geocoding, but no semantic reasoning could be made about any of the address components until it was at least partially resolved. Creating an address only containing a region_code and address_lines, and then geocoding is the recommended way to handle completely unstructured addresses (as opposed to guessing which parts of the address should be localities or administrative areas). "A String", ], @@ -208,8 +208,8 @@

Method Details

Returns: An object of the form: - { # Collection of information related to a business. - "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. + { # The `BusinessInfo` message contains essential information about a merchant's business. This message captures key business details such as physical address, customer service contacts, and region-specific identifiers. + "address": { # Represents a postal address, e.g. for postal delivery or payments addresses. Given a postal address, a postal service can deliver items to a premise, P.O. Box or similar. It is not intended to model geographical locations (roads, towns, mountains). In typical usage an address would be created via user input or from importing existing data, depending on the type of process. Advice on address input / editing: - Use an internationalization-ready address widget such as https://github.com/google/libaddressinput) - Users should not be presented with UI elements for input or editing of fields outside countries where that field is used. For more guidance on how to use this schema, please see: https://support.google.com/business/answer/6397478 # Optional. The address of the business. Only `region_code`, `address_lines`, `postal_code`, `administrative_area` and `locality` fields are supported. All other fields are ignored. "addressLines": [ # Unstructured address lines describing the lower levels of an address. Because values in address_lines do not have type information and may sometimes contain multiple values in a single field (e.g. "Austin, TX"), it is important that the line order is clear. The order of address lines should be "envelope order" for the country/region of the address. In places where this can vary (e.g. Japan), address_language is used to make it explicit (e.g. "ja" for large-to-small ordering and "ja-Latn" or "en" for small-to-large). This way, the most specific line of an address can be selected based on the language. The minimum permitted structural representation of an address consists of a region_code with all remaining information placed in the address_lines. It would be possible to format such an address very approximately without geocoding, but no semantic reasoning could be made about any of the address components until it was at least partially resolved. Creating an address only containing a region_code and address_lines, and then geocoding is the recommended way to handle completely unstructured addresses (as opposed to guessing which parts of the address should be localities or administrative areas). "A String", ], diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.homepage.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.homepage.html index 7bce9f12dfd..8aeea12f50b 100644 --- a/docs/dyn/merchantapi_accounts_v1beta.accounts.homepage.html +++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.homepage.html @@ -111,7 +111,7 @@

Method Details

Returns: An object of the form: - { # A store's homepage. + { # The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793). "claimed": True or False, # Output only. Whether the homepage is claimed. See https://support.google.com/merchants/answer/176793. "name": "A String", # Identifier. The resource name of the store's homepage. Format: `accounts/{account}/homepage` "uri": "A String", # Required. The URI (typically a URL) of the store's homepage. @@ -137,7 +137,7 @@

Method Details

Returns: An object of the form: - { # A store's homepage. + { # The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793). "claimed": True or False, # Output only. Whether the homepage is claimed. See https://support.google.com/merchants/answer/176793. "name": "A String", # Identifier. The resource name of the store's homepage. Format: `accounts/{account}/homepage` "uri": "A String", # Required. The URI (typically a URL) of the store's homepage. @@ -164,7 +164,7 @@

Method Details

Returns: An object of the form: - { # A store's homepage. + { # The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793). "claimed": True or False, # Output only. Whether the homepage is claimed. See https://support.google.com/merchants/answer/176793. "name": "A String", # Identifier. The resource name of the store's homepage. Format: `accounts/{account}/homepage` "uri": "A String", # Required. The URI (typically a URL) of the store's homepage. @@ -180,7 +180,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # A store's homepage. +{ # The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793). "claimed": True or False, # Output only. Whether the homepage is claimed. See https://support.google.com/merchants/answer/176793. "name": "A String", # Identifier. The resource name of the store's homepage. Format: `accounts/{account}/homepage` "uri": "A String", # Required. The URI (typically a URL) of the store's homepage. @@ -195,7 +195,7 @@

Method Details

Returns: An object of the form: - { # A store's homepage. + { # The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793). "claimed": True or False, # Output only. Whether the homepage is claimed. See https://support.google.com/merchants/answer/176793. "name": "A String", # Identifier. The resource name of the store's homepage. Format: `accounts/{account}/homepage` "uri": "A String", # Required. The URI (typically a URL) of the store's homepage. diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.html index 02d35614c5d..cd345377573 100644 --- a/docs/dyn/merchantapi_accounts_v1beta.accounts.html +++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.html @@ -176,7 +176,7 @@

Method Details

The object takes the form of: { # Request message for the `CreateAndConfigureAccount` method. - "account": { # An account. # Required. The account to be created. + "account": { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. # Required. The account to be created. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -198,7 +198,7 @@

Method Details

"users": [ # Optional. Users to be added to the account. { # Request message for the `CreateUser` method. "parent": "A String", # Required. The resource name of the account for which a user will be created. Format: `accounts/{account}` - "user": { # A [user](https://support.google.com/merchants/answer/12160472). # Required. The user to create. + "user": { # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). # Required. The user to create. "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -218,7 +218,7 @@

Method Details

Returns: An object of the form: - { # An account. + { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -265,7 +265,7 @@

Method Details

Returns: An object of the form: - { # An account. + { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -297,7 +297,7 @@

Method Details

{ # Response message for the `ListAccounts` method. "accounts": [ # The accounts matching the `ListAccountsRequest`. - { # An account. + { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -332,7 +332,7 @@

Method Details

{ # Response message for the `ListSubAccounts` method. "accounts": [ # The accounts for which the given parent account is an aggregator. - { # An account. + { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -386,7 +386,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # An account. +{ # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. @@ -408,7 +408,7 @@

Method Details

Returns: An object of the form: - { # An account. + { # The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on. "accountId": "A String", # Output only. The ID of the account. "accountName": "A String", # Required. A human-readable name of the account. See [store name](https://support.google.com/merchants/answer/160556) and [business name](https://support.google.com/merchants/answer/12159159) for more information. "adultContent": True or False, # Whether this account contains adult content. diff --git a/docs/dyn/merchantapi_accounts_v1beta.accounts.users.html b/docs/dyn/merchantapi_accounts_v1beta.accounts.users.html index e91553645fc..9ac73f15dd7 100644 --- a/docs/dyn/merchantapi_accounts_v1beta.accounts.users.html +++ b/docs/dyn/merchantapi_accounts_v1beta.accounts.users.html @@ -110,7 +110,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # A [user](https://support.google.com/merchants/answer/12160472). +{ # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -127,7 +127,7 @@

Method Details

Returns: An object of the form: - { # A [user](https://support.google.com/merchants/answer/12160472). + { # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -168,7 +168,7 @@

Method Details

Returns: An object of the form: - { # A [user](https://support.google.com/merchants/answer/12160472). + { # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -196,7 +196,7 @@

Method Details

{ # Response message for the `ListUsers` method. "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. "users": [ # The users from the specified account. - { # A [user](https://support.google.com/merchants/answer/12160472). + { # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -230,7 +230,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # A [user](https://support.google.com/merchants/answer/12160472). +{ # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], @@ -247,7 +247,7 @@

Method Details

Returns: An object of the form: - { # A [user](https://support.google.com/merchants/answer/12160472). + { # The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472). "accessRights": [ # Optional. The [access rights](https://support.google.com/merchants/answer/12160472?sjid=6789834943175119429-EU#accesstypes) the user has. "A String", ], diff --git a/docs/dyn/merchantapi_accounts_v1beta.termsOfService.html b/docs/dyn/merchantapi_accounts_v1beta.termsOfService.html index 41d6459af41..72972377750 100644 --- a/docs/dyn/merchantapi_accounts_v1beta.termsOfService.html +++ b/docs/dyn/merchantapi_accounts_v1beta.termsOfService.html @@ -126,7 +126,7 @@

Method Details

Returns: An object of the form: - { # A `TermsOfService`. + { # The `TermsOfService` message represents a specific version of the terms of service that merchants must accept to access certain features or services (see https://support.google.com/merchants/answer/160173). This message is important for the onboarding process, ensuring that merchants agree to the necessary legal agreements for using the service. Merchants can retrieve the latest terms of service for a given `kind` and `region` through `RetrieveLatestTermsOfService`, and accept them as required through `AcceptTermsOfService`. "external": True or False, # Whether this terms of service version is external. External terms of service versions can only be agreed through external processes and not directly by the merchant through UI or API. "fileUri": "A String", # URI for terms of service file that needs to be displayed to signing users. "kind": "A String", # The Kind this terms of service version applies to. @@ -153,7 +153,7 @@

Method Details

Returns: An object of the form: - { # A `TermsOfService`. + { # The `TermsOfService` message represents a specific version of the terms of service that merchants must accept to access certain features or services (see https://support.google.com/merchants/answer/160173). This message is important for the onboarding process, ensuring that merchants agree to the necessary legal agreements for using the service. Merchants can retrieve the latest terms of service for a given `kind` and `region` through `RetrieveLatestTermsOfService`, and accept them as required through `AcceptTermsOfService`. "external": True or False, # Whether this terms of service version is external. External terms of service versions can only be agreed through external processes and not directly by the merchant through UI or API. "fileUri": "A String", # URI for terms of service file that needs to be displayed to signing users. "kind": "A String", # The Kind this terms of service version applies to. diff --git a/docs/dyn/migrationcenter_v1alpha1.projects.locations.html b/docs/dyn/migrationcenter_v1alpha1.projects.locations.html index 3037f0de281..6f0a41b781d 100644 --- a/docs/dyn/migrationcenter_v1alpha1.projects.locations.html +++ b/docs/dyn/migrationcenter_v1alpha1.projects.locations.html @@ -190,6 +190,7 @@

Method Details

An object of the form: { # Describes the Migration Center settings related to the project. + "customerConsentForGoogleSalesToAccessMigrationCenter": True or False, # Customer consent for Google sales to access their Cloud Migration Center project. "disableCloudLogging": True or False, # Disable Cloud Logging for the Migration Center API. Users are billed for the logs. "name": "A String", # Output only. The name of the resource. "preferenceSet": "A String", # The preference set used by default for a project. @@ -255,6 +256,7 @@

Method Details

The object takes the form of: { # Describes the Migration Center settings related to the project. + "customerConsentForGoogleSalesToAccessMigrationCenter": True or False, # Customer consent for Google sales to access their Cloud Migration Center project. "disableCloudLogging": True or False, # Disable Cloud Logging for the Migration Center API. Users are billed for the logs. "name": "A String", # Output only. The name of the resource. "preferenceSet": "A String", # The preference set used by default for a project. diff --git a/docs/dyn/networkmanagement_v1.projects.locations.global_.connectivityTests.html b/docs/dyn/networkmanagement_v1.projects.locations.global_.connectivityTests.html index 34b98e228cb..3e1dd130b50 100644 --- a/docs/dyn/networkmanagement_v1.projects.locations.global_.connectivityTests.html +++ b/docs/dyn/networkmanagement_v1.projects.locations.global_.connectivityTests.html @@ -139,7 +139,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -503,7 +504,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -614,7 +616,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -978,7 +981,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1080,7 +1084,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1444,7 +1449,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1505,7 +1511,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1869,7 +1876,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. diff --git a/docs/dyn/networkmanagement_v1beta1.projects.locations.global_.connectivityTests.html b/docs/dyn/networkmanagement_v1beta1.projects.locations.global_.connectivityTests.html index c3cb66184ef..f9989051514 100644 --- a/docs/dyn/networkmanagement_v1beta1.projects.locations.global_.connectivityTests.html +++ b/docs/dyn/networkmanagement_v1beta1.projects.locations.global_.connectivityTests.html @@ -139,7 +139,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -503,7 +504,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -614,7 +616,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -978,7 +981,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1080,7 +1084,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1444,7 +1449,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1505,7 +1511,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. @@ -1869,7 +1876,8 @@

Method Details

"cloudSqlInstance": "A String", # A [Cloud SQL](https://cloud.google.com/sql) instance URI. "forwardingRule": "A String", # A forwarding rule and its corresponding IP address represent the frontend configuration of a Google Cloud load balancer. Forwarding rules are also used for protocol forwarding, Private Service Connect and other network services to provide forwarding information in the control plane. Format: projects/{project}/global/forwardingRules/{id} or projects/{project}/regions/{region}/forwardingRules/{id} "forwardingRuleTarget": "A String", # Output only. Specifies the type of the target of the forwarding rule. - "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). + "fqdn": "A String", # DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address. + "gkeMasterCluster": "A String", # A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). "instance": "A String", # A Compute Engine instance URI. "ipAddress": "A String", # The IP address of the endpoint, which can be an external or internal IP. "loadBalancerId": "A String", # Output only. ID of the load balancer the forwarding rule points to. Empty for forwarding rules not related to load balancers. diff --git a/docs/dyn/osconfig_v2beta.folders.locations.global_.policyOrchestrators.html b/docs/dyn/osconfig_v2beta.folders.locations.global_.policyOrchestrators.html index b61abe99dbd..c68ba4fc812 100644 --- a/docs/dyn/osconfig_v2beta.folders.locations.global_.policyOrchestrators.html +++ b/docs/dyn/osconfig_v2beta.folders.locations.global_.policyOrchestrators.html @@ -106,7 +106,7 @@

Method Details

Creates a new policy orchestrator under the given folder resource. `name` field of the given orchestrator are ignored and instead replaced by a product of `parent` and `policy_orchestrator_id`. Orchestrator state field might be only set to `ACTIVE`, `STOPPED` or omitted (in which case, the created resource will be in `ACTIVE` state anyway).
 
 Args:
-  parent: string, Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global' (required)
+  parent: string, Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
diff --git a/docs/dyn/osconfig_v2beta.organizations.locations.global_.policyOrchestrators.html b/docs/dyn/osconfig_v2beta.organizations.locations.global_.policyOrchestrators.html
index c8e57689fab..92845768a85 100644
--- a/docs/dyn/osconfig_v2beta.organizations.locations.global_.policyOrchestrators.html
+++ b/docs/dyn/osconfig_v2beta.organizations.locations.global_.policyOrchestrators.html
@@ -106,7 +106,7 @@ 

Method Details

Creates a new policy orchestrator under the given organizations resource. `name` field of the given orchestrator are ignored and instead replaced by a product of `parent` and `policy_orchestrator_id`. Orchestrator state field might be only set to `ACTIVE`, `STOPPED` or omitted (in which case, the created resource will be in `ACTIVE` state anyway).
 
 Args:
-  parent: string, Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global' (required)
+  parent: string, Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
diff --git a/docs/dyn/osconfig_v2beta.projects.locations.global_.policyOrchestrators.html b/docs/dyn/osconfig_v2beta.projects.locations.global_.policyOrchestrators.html
index f01387e15d8..d0145960a3d 100644
--- a/docs/dyn/osconfig_v2beta.projects.locations.global_.policyOrchestrators.html
+++ b/docs/dyn/osconfig_v2beta.projects.locations.global_.policyOrchestrators.html
@@ -106,7 +106,7 @@ 

Method Details

Creates a new policy orchestrator under the given project resource. `name` field of the given orchestrator are ignored and instead replaced by a product of `parent` and `policy_orchestrator_id`. Orchestrator state field might be only set to `ACTIVE`, `STOPPED` or omitted (in which case, the created resource will be in `ACTIVE` state anyway).
 
 Args:
-  parent: string, Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global' (required)
+  parent: string, Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global` (required)
   body: object, The request body.
     The object takes the form of:
 
diff --git a/docs/dyn/places_v1.places.html b/docs/dyn/places_v1.places.html
index b810ff690d9..c6bde4f9ffc 100644
--- a/docs/dyn/places_v1.places.html
+++ b/docs/dyn/places_v1.places.html
@@ -363,7 +363,7 @@ 

Method Details

"providerUri": "A String", # URI to the Place's data provider. }, ], - "businessStatus": "A String", # The business status for the place. + "businessStatus": "A String", "curbsidePickup": True or False, # Specifies if the business supports curbside pickup. "currentOpeningHours": { # Information about business hour of the place. # The hours of operation for the next seven days (including today). The time period starts at midnight on the date of the request and ends at 11:59 pm six days later. This field includes the special_days subfield of all hours, set for dates that have exceptional hours. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. @@ -896,7 +896,7 @@

Method Details

"providerUri": "A String", # URI to the Place's data provider. }, ], - "businessStatus": "A String", # The business status for the place. + "businessStatus": "A String", "curbsidePickup": True or False, # Specifies if the business supports curbside pickup. "currentOpeningHours": { # Information about business hour of the place. # The hours of operation for the next seven days (including today). The time period starts at midnight on the date of the request and ends at 11:59 pm six days later. This field includes the special_days subfield of all hours, set for dates that have exceptional hours. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. @@ -1558,7 +1558,7 @@

Method Details

"providerUri": "A String", # URI to the Place's data provider. }, ], - "businessStatus": "A String", # The business status for the place. + "businessStatus": "A String", "curbsidePickup": True or False, # Specifies if the business supports curbside pickup. "currentOpeningHours": { # Information about business hour of the place. # The hours of operation for the next seven days (including today). The time period starts at midnight on the date of the request and ends at 11:59 pm six days later. This field includes the special_days subfield of all hours, set for dates that have exceptional hours. "openNow": True or False, # Whether the opening hours period is currently active. For regular opening hours and current opening hours, this field means whether the place is open. For secondary opening hours and current secondary opening hours, this field means whether the secondary hours of this place is active. diff --git a/docs/dyn/recaptchaenterprise_v1.projects.assessments.html b/docs/dyn/recaptchaenterprise_v1.projects.assessments.html index 0b598a8c025..ff593a2453b 100644 --- a/docs/dyn/recaptchaenterprise_v1.projects.assessments.html +++ b/docs/dyn/recaptchaenterprise_v1.projects.assessments.html @@ -323,7 +323,7 @@

Method Details

"reencryptedUserCredentialsHash": "A String", # Output only. Corresponds to the re-encryption of the `encrypted_user_credentials_hash` field. It is used to match potential password leaks within `encrypted_leak_match_prefixes`. }, "riskAnalysis": { # Risk analysis result for an event. # Output only. The risk analysis result for the event being assessed. - "challenge": "A String", # Output only. Challenge information for SCORE_AND_CHALLENGE keys + "challenge": "A String", # Output only. Challenge information for SCORE_AND_CHALLENGE and INVISIBLE keys "extendedVerdictReasons": [ # Output only. Extended verdict reasons to be used for experimentation only. The set of possible reasons is subject to change. "A String", ], @@ -540,7 +540,7 @@

Method Details

"reencryptedUserCredentialsHash": "A String", # Output only. Corresponds to the re-encryption of the `encrypted_user_credentials_hash` field. It is used to match potential password leaks within `encrypted_leak_match_prefixes`. }, "riskAnalysis": { # Risk analysis result for an event. # Output only. The risk analysis result for the event being assessed. - "challenge": "A String", # Output only. Challenge information for SCORE_AND_CHALLENGE keys + "challenge": "A String", # Output only. Challenge information for SCORE_AND_CHALLENGE and INVISIBLE keys "extendedVerdictReasons": [ # Output only. Extended verdict reasons to be used for experimentation only. The set of possible reasons is subject to change. "A String", ], diff --git a/docs/dyn/recaptchaenterprise_v1.projects.keys.html b/docs/dyn/recaptchaenterprise_v1.projects.keys.html index 676631abeda..f0cc309b2fe 100644 --- a/docs/dyn/recaptchaenterprise_v1.projects.keys.html +++ b/docs/dyn/recaptchaenterprise_v1.projects.keys.html @@ -694,7 +694,7 @@

Method Details

body: object, The request body. The object takes the form of: -{ # The removeIpOverride request message. +{ # The RemoveIpOverride request message. "ipOverrideData": { # Information about the IP or IP range override. # Required. IP override to be removed from the key. "ip": "A String", # Required. The IP address to override (can be IPv4, IPv6 or CIDR). The IP override must be a valid IPv4 or IPv6 address, or a CIDR range. The IP override must be a public IP address. Example of IPv4: 168.192.5.6 Example of IPv6: 2001:0000:130F:0000:0000:09C0:876A:130B Example of IPv4 with CIDR: 168.192.5.0/24 Example of IPv6 with CIDR: 2001:0DB8:1234::/48 "overrideType": "A String", # Required. Describes the type of IP override. diff --git a/docs/dyn/tpu_v2alpha1.projects.locations.nodes.html b/docs/dyn/tpu_v2alpha1.projects.locations.nodes.html index be3afe56cb4..67643ddf5a7 100644 --- a/docs/dyn/tpu_v2alpha1.projects.locations.nodes.html +++ b/docs/dyn/tpu_v2alpha1.projects.locations.nodes.html @@ -98,6 +98,9 @@

Instance Methods

patch(name, body=None, updateMask=None, x__xgafv=None)

Updates the configurations of a node.

+

+ performMaintenance(name, body=None, x__xgafv=None)

+

Perform manual maintenance on a node.

simulateMaintenanceEvent(name, body=None, x__xgafv=None)

Simulates a maintenance event.

@@ -209,6 +212,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, } nodeId: string, The unqualified resource name. @@ -379,6 +390,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, }
@@ -530,6 +549,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, }, ], "unreachable": [ # Locations that could not be reached. @@ -648,6 +675,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, } updateMask: string, Required. Mask of fields from Node to update. Supported fields: [description, tags, labels, metadata, network_config.enable_external_ips]. @@ -680,6 +715,47 @@

Method Details

}
+
+ performMaintenance(name, body=None, x__xgafv=None) +
Perform manual maintenance on a node.
+
+Args:
+  name: string, Required. The resource name. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request for PerformMaintenance.
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
simulateMaintenanceEvent(name, body=None, x__xgafv=None)
Simulates a maintenance event.
diff --git a/docs/dyn/tpu_v2alpha1.projects.locations.queuedResources.html b/docs/dyn/tpu_v2alpha1.projects.locations.queuedResources.html
index d21ff4b80dd..0262210e947 100644
--- a/docs/dyn/tpu_v2alpha1.projects.locations.queuedResources.html
+++ b/docs/dyn/tpu_v2alpha1.projects.locations.queuedResources.html
@@ -92,6 +92,9 @@ 

Instance Methods

list_next()

Retrieves the next page of results.

+

+ performMaintenanceQueuedResource(name, body=None, x__xgafv=None)

+

Perform manual maintenance on specific nodes of a QueuedResource.

reset(name, body=None, x__xgafv=None)

Resets a QueuedResource TPU instance

@@ -256,6 +259,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, }, "nodeId": "A String", # The unqualified resource name. Should follow the `^[A-Za-z0-9_.~+%-]+$` regex format. This is only specified when requesting a single node. In case of multi-node requests, multi_node_params must be populated instead. It's an error to specify both node_id and multi_node_params. "parent": "A String", # Required. The parent resource name. @@ -492,6 +503,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, }, "nodeId": "A String", # The unqualified resource name. Should follow the `^[A-Za-z0-9_.~+%-]+$` regex format. This is only specified when requesting a single node. In case of multi-node requests, multi_node_params must be populated instead. It's an error to specify both node_id and multi_node_params. "parent": "A String", # Required. The parent resource name. @@ -666,6 +685,14 @@

Method Details

"tags": [ # Tags to apply to the TPU Node. Tags are used to identify valid sources or targets for network firewalls. "A String", ], + "upcomingMaintenance": { # Upcoming Maintenance notification information. # Output only. Upcoming maintenance on this TPU node. + "canReschedule": True or False, # Indicates if the maintenance can be customer triggered. + "latestWindowStartTime": "A String", # The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format. + "maintenanceStatus": "A String", # The status of the maintenance. + "type": "A String", # Defines the type of maintenance. + "windowEndTime": "A String", # The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format. + "windowStartTime": "A String", # The current start time of the maintenance window. This timestamp value is in RFC3339 text format. + }, }, "nodeId": "A String", # The unqualified resource name. Should follow the `^[A-Za-z0-9_.~+%-]+$` regex format. This is only specified when requesting a single node. In case of multi-node requests, multi_node_params must be populated instead. It's an error to specify both node_id and multi_node_params. "parent": "A String", # Required. The parent resource name. @@ -694,6 +721,50 @@

Method Details

+
+ performMaintenanceQueuedResource(name, body=None, x__xgafv=None) +
Perform manual maintenance on specific nodes of a QueuedResource.
+
+Args:
+  name: string, Required. The name of the QueuedResource which holds the nodes to perform maintenance on. (required)
+  body: object, The request body.
+    The object takes the form of:
+
+{ # Request for PerformMaintenanceQueuedResource.
+  "nodeNames": [ # The names of the nodes to perform maintenance on.
+    "A String",
+  ],
+}
+
+  x__xgafv: string, V1 error format.
+    Allowed values
+      1 - v1 error format
+      2 - v2 error format
+
+Returns:
+  An object of the form:
+
+    { # This resource represents a long-running operation that is the result of a network API call.
+  "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.
+  "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation.
+    "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+    "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use.
+      {
+        "a_key": "", # Properties of the object. Contains field @type with type URL.
+      },
+    ],
+    "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client.
+  },
+  "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+  "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`.
+  "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`.
+    "a_key": "", # Properties of the object. Contains field @type with type URL.
+  },
+}
+
+
reset(name, body=None, x__xgafv=None)
Resets a QueuedResource TPU instance
diff --git a/docs/dyn/youtube_v3.playlistItems.html b/docs/dyn/youtube_v3.playlistItems.html
index b2fcc386ccb..2731a0e184f 100644
--- a/docs/dyn/youtube_v3.playlistItems.html
+++ b/docs/dyn/youtube_v3.playlistItems.html
@@ -335,7 +335,7 @@ 

Method Details

}, }, ], - "kind": "youtube#playlistItemListResponse", # Identifies what kind of resource this is. Value: the fixed string "youtube#playlistItemListResponse". Etag of this resource. + "kind": "youtube#playlistItemListResponse", # Identifies what kind of resource this is. Value: the fixed string "youtube#playlistItemListResponse". "nextPageToken": "A String", # The token that can be used as the value of the pageToken parameter to retrieve the next page in the result set. "pageInfo": { # Paging details for lists of resources, including total number of items available and number of resources returned in a single page. # General pagination information. "resultsPerPage": 42, # The number of results included in the API response. diff --git a/docs/dyn/youtube_v3.videos.html b/docs/dyn/youtube_v3.videos.html index 9f5aeceb83c..12e2182fbfd 100644 --- a/docs/dyn/youtube_v3.videos.html +++ b/docs/dyn/youtube_v3.videos.html @@ -405,6 +405,7 @@

Method Details

"viewCount": "A String", # The number of times the video has been viewed. }, "status": { # Basic details about a video category, such as its localized title. Next Id: 19 # The status object contains information about the video's uploading, processing, and privacy statuses. + "containsSyntheticMedia": True or False, # Indicates if the video contains altered or synthetic media. "embeddable": True or False, # This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update "failureReason": "A String", # This value explains why a video failed to upload. This property is only present if the uploadStatus property indicates that the upload failed. "license": "A String", # The video's license. @mutable youtube.videos.insert youtube.videos.update @@ -712,6 +713,7 @@

Method Details

"viewCount": "A String", # The number of times the video has been viewed. }, "status": { # Basic details about a video category, such as its localized title. Next Id: 19 # The status object contains information about the video's uploading, processing, and privacy statuses. + "containsSyntheticMedia": True or False, # Indicates if the video contains altered or synthetic media. "embeddable": True or False, # This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update "failureReason": "A String", # This value explains why a video failed to upload. This property is only present if the uploadStatus property indicates that the upload failed. "license": "A String", # The video's license. @mutable youtube.videos.insert youtube.videos.update @@ -1042,6 +1044,7 @@

Method Details

"viewCount": "A String", # The number of times the video has been viewed. }, "status": { # Basic details about a video category, such as its localized title. Next Id: 19 # The status object contains information about the video's uploading, processing, and privacy statuses. + "containsSyntheticMedia": True or False, # Indicates if the video contains altered or synthetic media. "embeddable": True or False, # This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update "failureReason": "A String", # This value explains why a video failed to upload. This property is only present if the uploadStatus property indicates that the upload failed. "license": "A String", # The video's license. @mutable youtube.videos.insert youtube.videos.update @@ -1412,6 +1415,7 @@

Method Details

"viewCount": "A String", # The number of times the video has been viewed. }, "status": { # Basic details about a video category, such as its localized title. Next Id: 19 # The status object contains information about the video's uploading, processing, and privacy statuses. + "containsSyntheticMedia": True or False, # Indicates if the video contains altered or synthetic media. "embeddable": True or False, # This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update "failureReason": "A String", # This value explains why a video failed to upload. This property is only present if the uploadStatus property indicates that the upload failed. "license": "A String", # The video's license. @mutable youtube.videos.insert youtube.videos.update @@ -1713,6 +1717,7 @@

Method Details

"viewCount": "A String", # The number of times the video has been viewed. }, "status": { # Basic details about a video category, such as its localized title. Next Id: 19 # The status object contains information about the video's uploading, processing, and privacy statuses. + "containsSyntheticMedia": True or False, # Indicates if the video contains altered or synthetic media. "embeddable": True or False, # This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update "failureReason": "A String", # This value explains why a video failed to upload. This property is only present if the uploadStatus property indicates that the upload failed. "license": "A String", # The video's license. @mutable youtube.videos.insert youtube.videos.update diff --git a/googleapiclient/discovery_cache/documents/accesscontextmanager.v1.json b/googleapiclient/discovery_cache/documents/accesscontextmanager.v1.json index ed851ef84df..b98e6ec450b 100644 --- a/googleapiclient/discovery_cache/documents/accesscontextmanager.v1.json +++ b/googleapiclient/discovery_cache/documents/accesscontextmanager.v1.json @@ -1206,7 +1206,7 @@ ], "parameters": { "append": { -"description": "Optional. This field controls whether or not certain repeated settings in the update request overwrite or append to existing settings on the binding. If true, then append. Otherwise overwrite. So far, only scoped_access_settings supports appending. Global access_levels, dry_run_access_levels, reauth_settings, and session_settings are not compatible with append functionality, and the request will return an error if append=true when these settings are in the update_mask. The request will also return an error if append=true when \"scoped_access_settings\" is not set in the update_mask.", +"description": "Optional. This field controls whether or not certain repeated settings in the update request overwrite or append to existing settings on the binding. If true, then append. Otherwise overwrite. So far, only scoped_access_settings with reauth_settings supports appending. Global access_levels, access_levels in scoped_access_settings, dry_run_access_levels, reauth_settings, and session_settings are not compatible with append functionality, and the request will return an error if append=true when these settings are in the update_mask. The request will also return an error if append=true when \"scoped_access_settings\" is not set in the update_mask.", "location": "query", "type": "boolean" }, @@ -1295,7 +1295,7 @@ } } }, -"revision": "20241022", +"revision": "20241028", "rootUrl": "https://accesscontextmanager.googleapis.com/", "schemas": { "AccessContextManagerOperationMetadata": { @@ -1884,7 +1884,7 @@ "type": "object" }, "GcpUserAccessBinding": { -"description": "Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access. Next ID: 11", +"description": "Restricts access to Cloud Console and Google Cloud APIs for a set of users using Context-Aware Access.", "id": "GcpUserAccessBinding", "properties": { "accessLevels": { @@ -1926,6 +1926,10 @@ "$ref": "ScopedAccessSettings" }, "type": "array" +}, +"sessionSettings": { +"$ref": "SessionSettings", +"description": "Optional. GCSL policy for the group key. Migrated from ReauthSettings" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/admin.directory_v1.json b/googleapiclient/discovery_cache/documents/admin.directory_v1.json index 030f10c5507..092434dfcc0 100644 --- a/googleapiclient/discovery_cache/documents/admin.directory_v1.json +++ b/googleapiclient/discovery_cache/documents/admin.directory_v1.json @@ -4671,7 +4671,7 @@ } } }, -"revision": "20240924", +"revision": "20241029", "rootUrl": "https://admin.googleapis.com/", "schemas": { "Alias": { @@ -5184,6 +5184,25 @@ false }, "type": "object" }, +"ByteUsage": { +"description": "Represents a data capacity with some amount of current usage in bytes.", +"id": "ByteUsage", +"properties": { +"capacityBytes": { +"description": "Output only. The total capacity value, in bytes.", +"format": "int64", +"readOnly": true, +"type": "string" +}, +"usedBytes": { +"description": "Output only. The current usage value, in bytes.", +"format": "int64", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, "CalendarResource": { "description": "Public API: Resources.calendars", "id": "CalendarResource", @@ -5640,6 +5659,11 @@ false "readOnly": true, "type": "string" }, +"diskSpaceUsage": { +"$ref": "ByteUsage", +"description": "Output only. How much disk space the device has available and is currently using.", +"readOnly": true +}, "diskVolumeReports": { "description": "Reports of disk space and other info about mounted/connected volumes.", "items": { diff --git a/googleapiclient/discovery_cache/documents/aiplatform.v1.json b/googleapiclient/discovery_cache/documents/aiplatform.v1.json index 39927321f34..cb03cd81e2c 100644 --- a/googleapiclient/discovery_cache/documents/aiplatform.v1.json +++ b/googleapiclient/discovery_cache/documents/aiplatform.v1.json @@ -791,6 +791,61 @@ } }, "projects": { +"methods": { +"getCacheConfig": { +"description": "Gets a GenAI cache config.", +"flatPath": "v1/projects/{projectsId}/cacheConfig", +"httpMethod": "GET", +"id": "aiplatform.projects.getCacheConfig", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. Name of the cache config. Format: - `projects/{project}/cacheConfig`.", +"location": "path", +"pattern": "^projects/[^/]+/cacheConfig$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"response": { +"$ref": "GoogleCloudAiplatformV1CacheConfig" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"updateCacheConfig": { +"description": "Updates a cache config.", +"flatPath": "v1/projects/{projectsId}/cacheConfig", +"httpMethod": "PATCH", +"id": "aiplatform.projects.updateCacheConfig", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Identifier. Name of the cache config. Format: - `projects/{project}/cacheConfig`.", +"location": "path", +"pattern": "^projects/[^/]+/cacheConfig$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+name}", +"request": { +"$ref": "GoogleCloudAiplatformV1CacheConfig" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +}, "resources": { "locations": { "methods": { @@ -3763,6 +3818,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"fetchPredictOperation": { +"description": "Fetch an asynchronous online prediction operation.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:fetchPredictOperation", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.endpoints.fetchPredictOperation", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/endpoints/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+endpoint}:fetchPredictOperation", +"request": { +"$ref": "GoogleCloudAiplatformV1FetchPredictOperationRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "generateContent": { "description": "Generate content with multimodal inputs.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:generateContent", @@ -3960,6 +4044,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"predictLongRunning": { +"description": "", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:predictLongRunning", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.endpoints.predictLongRunning", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/endpoints/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+endpoint}:predictLongRunning", +"request": { +"$ref": "GoogleCloudAiplatformV1PredictLongRunningRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "rawPredict": { "description": "Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:rawPredict", @@ -13932,6 +14045,35 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, +"fetchPredictOperation": { +"description": "Fetch an asynchronous online prediction operation.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:fetchPredictOperation", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.publishers.models.fetchPredictOperation", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/publishers/[^/]+/models/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+endpoint}:fetchPredictOperation", +"request": { +"$ref": "GoogleCloudAiplatformV1FetchPredictOperationRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "generateContent": { "description": "Generate content with multimodal inputs.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:generateContent", @@ -13990,6 +14132,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"predictLongRunning": { +"description": "", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:predictLongRunning", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.publishers.models.predictLongRunning", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/publishers/[^/]+/models/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+endpoint}:predictLongRunning", +"request": { +"$ref": "GoogleCloudAiplatformV1PredictLongRunningRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "rawPredict": { "description": "Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:rawPredict", @@ -17902,7 +18073,7 @@ } } }, -"revision": "20241025", +"revision": "20241026", "rootUrl": "https://aiplatform.googleapis.com/", "schemas": { "CloudAiLargeModelsVisionGenerateVideoResponse": { @@ -18198,6 +18369,15 @@ "description": "Video", "id": "CloudAiLargeModelsVisionVideo", "properties": { +"encodedVideo": { +"description": "Base 64 encoded video bytes.", +"format": "byte", +"type": "string" +}, +"encoding": { +"description": "Video encoding, for example \"video/mp4\".", +"type": "string" +}, "uri": { "description": "Path to another storage (typically Google Cloud Storage).", "type": "string" @@ -19464,6 +19644,21 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1CacheConfig": { +"description": "Config of GenAI caching features. This is a singleton resource.", +"id": "GoogleCloudAiplatformV1CacheConfig", +"properties": { +"disableCache": { +"description": "If set to true, disables GenAI caching. Otherwise caching is enabled.", +"type": "boolean" +}, +"name": { +"description": "Identifier. Name of the cache config. Format: - `projects/{project}/cacheConfig`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1CancelBatchPredictionJobRequest": { "description": "Request message for JobService.CancelBatchPredictionJob.", "id": "GoogleCloudAiplatformV1CancelBatchPredictionJobRequest", @@ -19556,7 +19751,7 @@ "Token generation reached a natural stopping point or a configured stop sequence.", "Token generation reached the configured maximum output tokens.", "Token generation stopped because the content potentially contains safety violations. NOTE: When streaming, content is empty if content filters blocks the output.", -"Token generation stopped because the content potentially contains copyright violations.", +"The token generation stopped because of potential recitation.", "All other reasons that stopped the token generation.", "Token generation stopped because the content contains forbidden terms.", "Token generation stopped for potentially containing prohibited content.", @@ -24290,6 +24485,17 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1FetchPredictOperationRequest": { +"description": "Request message for PredictionService.FetchPredictOperation.", +"id": "GoogleCloudAiplatformV1FetchPredictOperationRequest", +"properties": { +"operationName": { +"description": "Required. The server-assigned name for the operation.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1FileData": { "description": "URI based data.", "id": "GoogleCloudAiplatformV1FileData", @@ -24646,7 +24852,7 @@ "type": "object" }, "GoogleCloudAiplatformV1FunctionDeclaration": { -"description": "Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client.", +"description": "Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client.", "id": "GoogleCloudAiplatformV1FunctionDeclaration", "properties": { "description": { @@ -29426,6 +29632,10 @@ false "readOnly": true, "type": "string" }, +"customEnvironmentSpec": { +"$ref": "GoogleCloudAiplatformV1NotebookExecutionJobCustomEnvironmentSpec", +"description": "The custom compute configuration for an execution job." +}, "dataformRepositorySource": { "$ref": "GoogleCloudAiplatformV1NotebookExecutionJobDataformRepositorySource", "description": "The Dataform Repository pointing to a single file notebook repository." @@ -29531,6 +29741,25 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1NotebookExecutionJobCustomEnvironmentSpec": { +"description": "Compute configuration to use for an execution job.", +"id": "GoogleCloudAiplatformV1NotebookExecutionJobCustomEnvironmentSpec", +"properties": { +"machineSpec": { +"$ref": "GoogleCloudAiplatformV1MachineSpec", +"description": "The specification of a single machine for the execution job." +}, +"networkSpec": { +"$ref": "GoogleCloudAiplatformV1NetworkSpec", +"description": "The network configuration to use for the execution job." +}, +"persistentDiskSpec": { +"$ref": "GoogleCloudAiplatformV1PersistentDiskSpec", +"description": "The specification of a persistent disk to attach for the execution job." +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1NotebookExecutionJobDataformRepositorySource": { "description": "The Dataform Repository containing the input notebook.", "id": "GoogleCloudAiplatformV1NotebookExecutionJobDataformRepositorySource", @@ -30838,6 +31067,24 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1PredictLongRunningRequest": { +"description": "Request message for PredictionService.PredictLongRunning.", +"id": "GoogleCloudAiplatformV1PredictLongRunningRequest", +"properties": { +"instances": { +"description": "Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.", +"items": { +"type": "any" +}, +"type": "array" +}, +"parameters": { +"description": "Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.", +"type": "any" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1PredictRequest": { "description": "Request message for PredictionService.Predict.", "id": "GoogleCloudAiplatformV1PredictRequest", @@ -35049,6 +35296,10 @@ false "structuredPrompt": { "$ref": "GoogleCloudAiplatformV1SchemaPromptSpecStructuredPrompt", "description": "The prompt variation that stores preambles in separate fields." +}, +"translationPrompt": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationPrompt", +"description": "The prompt variation for Translation use case." } }, "type": "object" @@ -35147,6 +35398,35 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePair": { +"description": "A pair of sentences used as reference in source and target languages.", +"id": "GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePair", +"properties": { +"sourceSentence": { +"description": "Source sentence in the sentence pair.", +"type": "string" +}, +"targetSentence": { +"description": "Target sentence in the sentence pair.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePairList": { +"description": "A list of reference sentence pairs.", +"id": "GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePairList", +"properties": { +"referenceSentencePairs": { +"description": "Reference sentence pairs.", +"items": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePair" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1SchemaPromptSpecStructuredPrompt": { "description": "Prompt variation that stores preambles in separate fields.", "id": "GoogleCloudAiplatformV1SchemaPromptSpecStructuredPrompt", @@ -35198,6 +35478,108 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationExample": { +"description": "The translation example that contains reference sentences from various sources.", +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationExample", +"properties": { +"referenceSentencePairLists": { +"description": "The reference sentences from inline text.", +"items": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecReferenceSentencePairList" +}, +"type": "array" +}, +"referenceSentencesFileInputs": { +"description": "The reference sentences from file.", +"items": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationSentenceFileInput" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationFileInputSource": { +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationFileInputSource", +"properties": { +"content": { +"description": "The file's contents.", +"type": "string" +}, +"displayName": { +"description": "The file's display name.", +"type": "string" +}, +"mimeType": { +"description": "The file's mime type.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationGcsInputSource": { +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationGcsInputSource", +"properties": { +"inputUri": { +"description": "Source data URI. For example, `gs://my_bucket/my_object`.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationOption": { +"description": "Optional settings for translation prompt.", +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationOption", +"properties": { +"numberOfShots": { +"description": "How many shots to use.", +"format": "int32", +"type": "integer" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationPrompt": { +"description": "Prompt variation for Translation use case.", +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationPrompt", +"properties": { +"example": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationExample", +"description": "The translation example." +}, +"option": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationOption", +"description": "The translation option." +}, +"promptMessage": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecPromptMessage", +"description": "The prompt message." +}, +"sourceLanguageCode": { +"description": "The source language code.", +"type": "string" +}, +"targetLanguageCode": { +"description": "The target language code.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1SchemaPromptSpecTranslationSentenceFileInput": { +"id": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationSentenceFileInput", +"properties": { +"fileInputSource": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationFileInputSource", +"description": "Inlined file source." +}, +"gcsInputSource": { +"$ref": "GoogleCloudAiplatformV1SchemaPromptSpecTranslationGcsInputSource", +"description": "Cloud Storage file source." +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1SchemaTablesDatasetMetadata": { "description": "The metadata of Datasets that contain tables data.", "id": "GoogleCloudAiplatformV1SchemaTablesDatasetMetadata", @@ -39907,7 +40289,7 @@ false "id": "GoogleCloudAiplatformV1TuningJob", "properties": { "baseModel": { -"description": "The base model that is being tuned, e.g., \"gemini-1.0-pro-002\".", +"description": "The base model that is being tuned, e.g., \"gemini-1.0-pro-002\". .", "type": "string" }, "createTime": { diff --git a/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json b/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json index 159548b4d55..ec9763a41d7 100644 --- a/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/aiplatform.v1beta1.json @@ -4565,6 +4565,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"fetchPredictOperation": { +"description": "Fetch an asynchronous online prediction operation.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:fetchPredictOperation", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.endpoints.fetchPredictOperation", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/endpoints/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+endpoint}:fetchPredictOperation", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1FetchPredictOperationRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "generateContent": { "description": "Generate content with multimodal inputs.", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:generateContent", @@ -4788,6 +4817,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"predictLongRunning": { +"description": "", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:predictLongRunning", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.endpoints.predictLongRunning", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/endpoints/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+endpoint}:predictLongRunning", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1PredictLongRunningRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "rawPredict": { "description": "Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/endpoints/{endpointsId}:rawPredict", @@ -6191,6 +6249,250 @@ } }, "resources": { +"featureMonitors": { +"methods": { +"create": { +"description": "Creates a new FeatureMonitor in a given project, location and FeatureGroup.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"featureMonitorId": { +"description": "Required. The ID to use for this FeatureMonitor, which will become the final component of the FeatureGroup's resource name. This value may be up to 60 characters, and valid characters are `[a-z0-9_]`. The first character cannot be a number. The value must be unique within the FeatureGroup.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The resource name of FeatureGroup to create FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featuregroup}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+parent}/featureMonitors", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitor" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"delete": { +"description": "Deletes a single FeatureMonitor.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors/{featureMonitorsId}", +"httpMethod": "DELETE", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the FeatureMonitor to be deleted. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+/featureMonitors/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"get": { +"description": "Gets details of a single FeatureMonitor.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors/{featureMonitorsId}", +"httpMethod": "GET", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the FeatureMonitor resource.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+/featureMonitors/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+name}", +"response": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitor" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"list": { +"description": "Lists FeatureGroups in a given project and location.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors", +"httpMethod": "GET", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.list", +"parameterOrder": [ +"parent" +], +"parameters": { +"filter": { +"description": "Optional. Lists the FeatureMonitors that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `update_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be in RFC 3339 format. * `labels`: Supports key-value equality and key presence. Examples: * `create_time > \"2020-01-01\" OR update_time > \"2020-01-01\"` FeatureMonitors created or updated after 2020-01-01. * `labels.env = \"prod\"` FeatureGroups with label \"env\" set to \"prod\".", +"location": "query", +"type": "string" +}, +"orderBy": { +"description": "Optional. A comma-separated list of fields to order by, sorted in ascending order. Use \"desc\" after a field name for descending. Supported Fields: * `create_time` * `update_time`", +"location": "query", +"type": "string" +}, +"pageSize": { +"description": "Optional. The maximum number of FeatureGroups to return. The service may return fewer than this value. If unspecified, at most 100 FeatureMonitors will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"pageToken": { +"description": "Optional. A page token, received from a previous FeatureRegistryService.ListFeatureMonitors call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to FeatureRegistryService.ListFeatureMonitors must match the call that provided the page token.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The resource name of the FeatureGroup to list FeatureMonitors. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+parent}/featureMonitors", +"response": { +"$ref": "GoogleCloudAiplatformV1beta1ListFeatureMonitorsResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +}, +"resources": { +"featureMonitorJobs": { +"methods": { +"create": { +"description": "Creates a new feature monitor job.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors/{featureMonitorsId}/featureMonitorJobs", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"featureMonitorJobId": { +"description": "Optional. Output only. System-generated ID for feature monitor job.", +"format": "int64", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The resource name of FeatureMonitor to create FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+/featureMonitors/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+parent}/featureMonitorJobs", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitorJob" +}, +"response": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitorJob" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"get": { +"description": "Get a feature monitor job.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors/{featureMonitorsId}/featureMonitorJobs/{featureMonitorJobsId}", +"httpMethod": "GET", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the FeatureMonitorJob resource. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+/featureMonitors/[^/]+/featureMonitorJobs/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+name}", +"response": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitorJob" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"list": { +"description": "List feature monitor jobs.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/featureGroups/{featureGroupsId}/featureMonitors/{featureMonitorsId}/featureMonitorJobs", +"httpMethod": "GET", +"id": "aiplatform.projects.locations.featureGroups.featureMonitors.featureMonitorJobs.list", +"parameterOrder": [ +"parent" +], +"parameters": { +"filter": { +"description": "Optional. Lists the FeatureMonitorJobs that match the filter expression. The following fields are supported: * `create_time`: Supports `=`, `!=`, `<`, `>`, `<=`, and `>=` comparisons. Values must be Examples: * `create_time > \"2020-01-01\"` FeatureMonitorJobs created after 2020-01-01.", +"location": "query", +"type": "string" +}, +"orderBy": { +"description": "Optional. A comma-separated list of fields to order by, sorted in ascending order. Use \"desc\" after a field name for descending. Supported Fields: * `create_time`", +"location": "query", +"type": "string" +}, +"pageSize": { +"description": "Optional. The maximum number of FeatureMonitorJobs to return. The service may return fewer than this value. If unspecified, at most 100 FeatureMonitorJobs will be returned. The maximum value is 100; any value greater than 100 will be coerced to 100.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"pageToken": { +"description": "Optional. A page token, received from a previous FeatureRegistryService.ListFeatureMonitorJobs call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to FeatureRegistryService.ListFeatureMonitorJobs must match the call that provided the page token.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The resource name of the FeatureMonitor to list FeatureMonitorJobs. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/featureGroups/[^/]+/featureMonitors/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+parent}/featureMonitorJobs", +"response": { +"$ref": "GoogleCloudAiplatformV1beta1ListFeatureMonitorJobsResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +} +} +}, "features": { "methods": { "batchCreate": { @@ -6288,6 +6590,24 @@ "name" ], "parameters": { +"featureStatsAndAnomalySpec.latestStatsCount": { +"description": "Optional. If set, returns the most recent count of stats. Valid value is [0, 100]. If stats_time_range is set, return most recent count of stats within the stats_time_range.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"featureStatsAndAnomalySpec.statsTimeRange.endTime": { +"description": "Optional. Exclusive end of the interval. If specified, a Timestamp matching this interval will have to be before the end.", +"format": "google-datetime", +"location": "query", +"type": "string" +}, +"featureStatsAndAnomalySpec.statsTimeRange.startTime": { +"description": "Optional. Inclusive start of the interval. If specified, a Timestamp matching this interval will have to be the same or after the start.", +"format": "google-datetime", +"location": "query", +"type": "string" +}, "name": { "description": "Required. The name of the Feature resource. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}`", "location": "path", @@ -8485,6 +8805,24 @@ "name" ], "parameters": { +"featureStatsAndAnomalySpec.latestStatsCount": { +"description": "Optional. If set, returns the most recent count of stats. Valid value is [0, 100]. If stats_time_range is set, return most recent count of stats within the stats_time_range.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"featureStatsAndAnomalySpec.statsTimeRange.endTime": { +"description": "Optional. Exclusive end of the interval. If specified, a Timestamp matching this interval will have to be before the end.", +"format": "google-datetime", +"location": "query", +"type": "string" +}, +"featureStatsAndAnomalySpec.statsTimeRange.startTime": { +"description": "Optional. Inclusive start of the interval. If specified, a Timestamp matching this interval will have to be the same or after the start.", +"format": "google-datetime", +"location": "query", +"type": "string" +}, "name": { "description": "Required. The name of the Feature resource. Format for entity_type as parent: `projects/{project}/locations/{location}/featurestores/{featurestore}/entityTypes/{entity_type}` Format for feature_group as parent: `projects/{project}/locations/{location}/featureGroups/{feature_group}`", "location": "path", @@ -16305,6 +16643,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"fetchPredictOperation": { +"description": "Fetch an asynchronous online prediction operation.", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:fetchPredictOperation", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.publishers.models.fetchPredictOperation", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/publishers/[^/]+/models/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+endpoint}:fetchPredictOperation", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1FetchPredictOperationRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "generateContent": { "description": "Generate content with multimodal inputs.", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:generateContent", @@ -16394,6 +16761,35 @@ "https://www.googleapis.com/auth/cloud-platform.read-only" ] }, +"predictLongRunning": { +"description": "", +"flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:predictLongRunning", +"httpMethod": "POST", +"id": "aiplatform.projects.locations.publishers.models.predictLongRunning", +"parameterOrder": [ +"endpoint" +], +"parameters": { +"endpoint": { +"description": "Required. The name of the Endpoint requested to serve the prediction. Format: `projects/{project}/locations/{location}/endpoints/{endpoint}` or `projects/{project}/locations/{location}/publishers/{publisher}/models/{model}`", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/publishers/[^/]+/models/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta1/{+endpoint}:predictLongRunning", +"request": { +"$ref": "GoogleCloudAiplatformV1beta1PredictLongRunningRequest" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-platform.read-only" +] +}, "rawPredict": { "description": "Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers: * `X-Vertex-AI-Endpoint-Id`: ID of the Endpoint that served this prediction. * `X-Vertex-AI-Deployed-Model-Id`: ID of the Endpoint's DeployedModel that served this prediction.", "flatPath": "v1beta1/projects/{projectsId}/locations/{locationsId}/publishers/{publishersId}/models/{modelsId}:rawPredict", @@ -21286,7 +21682,7 @@ } } }, -"revision": "20241025", +"revision": "20241026", "rootUrl": "https://aiplatform.googleapis.com/", "schemas": { "CloudAiLargeModelsVisionGenerateVideoResponse": { @@ -21582,6 +21978,15 @@ "description": "Video", "id": "CloudAiLargeModelsVisionVideo", "properties": { +"encodedVideo": { +"description": "Base 64 encoded video bytes.", +"format": "byte", +"type": "string" +}, +"encoding": { +"description": "Video encoding, for example \"video/mp4\".", +"type": "string" +}, "uri": { "description": "Path to another storage (typically Google Cloud Storage).", "type": "string" @@ -23339,7 +23744,7 @@ "Token generation reached a natural stopping point or a configured stop sequence.", "Token generation reached the configured maximum output tokens.", "Token generation stopped because the content potentially contains safety violations. NOTE: When streaming, content is empty if content filters blocks the output.", -"Token generation stopped because the content potentially contains copyright violations.", +"The token generation stopped because of potential recitation.", "All other reasons that stopped the token generation.", "Token generation stopped because the content contains forbidden terms.", "Token generation stopped for potentially containing prohibited content.", @@ -25488,6 +25893,10 @@ "$ref": "GoogleCloudAiplatformV1beta1ExplanationSpec", "description": "Explanation configuration for this DeployedModel. When deploying a Model using EndpointService.DeployModel, this value overrides the value of Model.explanation_spec. All fields of explanation_spec are optional in the request. If a field of explanation_spec is not populated, the value of the same field of Model.explanation_spec is inherited. If the corresponding Model.explanation_spec is not populated, all fields of the explanation_spec will be used for the explanation configuration." }, +"fasterDeploymentConfig": { +"$ref": "GoogleCloudAiplatformV1beta1FasterDeploymentConfig", +"description": "Configuration for faster model deployment." +}, "id": { "description": "Immutable. The ID of the DeployedModel. If not provided upon deployment, Vertex AI will generate a value for this ID. This value should be 1-10 characters, and valid characters are `/[0-9]/`.", "type": "string" @@ -27581,6 +27990,17 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1FasterDeploymentConfig": { +"description": "Configuration for faster model deployment.", +"id": "GoogleCloudAiplatformV1beta1FasterDeploymentConfig", +"properties": { +"fastTryoutEnabled": { +"description": "If true, enable fast tryout feature for this deployed model.", +"type": "boolean" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1Feature": { "description": "Feature Metadata information. For example, color is a feature that describes an apple.", "id": "GoogleCloudAiplatformV1beta1Feature", @@ -27603,6 +28023,14 @@ "description": "Used to perform a consistent read-modify-write updates. If not set, a blind \"overwrite\" update happens.", "type": "string" }, +"featureStatsAndAnomaly": { +"description": "Output only. Only applicable for Vertex AI Feature Store. The list of historical stats and anomalies.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureStatsAndAnomaly" +}, +"readOnly": true, +"type": "array" +}, "labels": { "additionalProperties": { "type": "string" @@ -27764,6 +28192,143 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1FeatureMonitor": { +"description": "Vertex AI Feature Monitor.", +"id": "GoogleCloudAiplatformV1beta1FeatureMonitor", +"properties": { +"createTime": { +"description": "Output only. Timestamp when this FeatureMonitor was created.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"description": { +"description": "Optional. Description of the FeatureMonitor.", +"type": "string" +}, +"etag": { +"description": "Optional. Used to perform consistent read-modify-write updates. If not set, a blind \"overwrite\" update happens.", +"type": "string" +}, +"featureSelectionConfig": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureSelectionConfig", +"description": "Required. Feature selection config for the FeatureMonitor." +}, +"labels": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. The labels with user-defined metadata to organize your FeatureMonitor. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded).\" System reserved label keys are prefixed with \"aiplatform.googleapis.com/\" and are immutable.", +"type": "object" +}, +"name": { +"description": "Identifier. Name of the FeatureMonitor. Format: `projects/{project}/locations/{location}/featureGroups/{featureGroup}/featureMonitors/{featureMonitor}`", +"type": "string" +}, +"scheduleConfig": { +"$ref": "GoogleCloudAiplatformV1beta1ScheduleConfig", +"description": "Required. Schedule config for the FeatureMonitor." +}, +"updateTime": { +"description": "Output only. Timestamp when this FeatureMonitor was last updated.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1FeatureMonitorJob": { +"description": "Vertex AI Feature Monitor Job.", +"id": "GoogleCloudAiplatformV1beta1FeatureMonitorJob", +"properties": { +"createTime": { +"description": "Output only. Timestamp when this FeatureMonitorJob was created. Creation of a FeatureMonitorJob means that the job is pending / waiting for sufficient resources but may not have started running yet.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"description": { +"description": "Optional. Description of the FeatureMonitor.", +"type": "string" +}, +"driftBaseFeatureMonitorJobId": { +"description": "Output only. FeatureMonitorJob ID comparing to which the drift is calculated.", +"format": "int64", +"readOnly": true, +"type": "string" +}, +"driftBaseSnapshotTime": { +"description": "Output only. Data snapshot time comparing to which the drift is calculated.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"featureSelectionConfig": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureSelectionConfig", +"description": "Output only. Feature selection config used when creating FeatureMonitorJob.", +"readOnly": true +}, +"finalStatus": { +"$ref": "GoogleRpcStatus", +"description": "Output only. Final status of the FeatureMonitorJob.", +"readOnly": true +}, +"jobSummary": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitorJobJobSummary", +"description": "Output only. Summary from the FeatureMonitorJob.", +"readOnly": true +}, +"labels": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. The labels with user-defined metadata to organize your FeatureMonitorJob. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information on and examples of labels. No more than 64 user labels can be associated with one FeatureMonitor(System labels are excluded).\" System reserved label keys are prefixed with \"aiplatform.googleapis.com/\" and are immutable.", +"type": "object" +}, +"name": { +"description": "Identifier. Name of the FeatureMonitorJob. Format: `projects/{project}/locations/{location}/featureGroups/{feature_group}/featureMonitors/{feature_monitor}/featureMonitorJobs/{feature_monitor_job}`.", +"type": "string" +}, +"triggerType": { +"description": "Output only. Trigger type of the Feature Monitor Job.", +"enum": [ +"FEATURE_MONITOR_JOB_TRIGGER_UNSPECIFIED", +"FEATURE_MONITOR_JOB_TRIGGER_PERIODIC", +"FEATURE_MONITOR_JOB_TRIGGER_ON_DEMAND" +], +"enumDescriptions": [ +"Trigger type unspecified.", +"Triggered by periodic schedule.", +"Triggered on demand by CreateFeatureMonitorJob request." +], +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1FeatureMonitorJobJobSummary": { +"description": "Summary from the FeatureMonitorJob.", +"id": "GoogleCloudAiplatformV1beta1FeatureMonitorJobJobSummary", +"properties": { +"featureStatsAndAnomalies": { +"description": "Output only. Features and their stats and anomalies", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureStatsAndAnomaly" +}, +"readOnly": true, +"type": "array" +}, +"totalSlotMs": { +"description": "Output only. BigQuery slot milliseconds consumed.", +"format": "int64", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1FeatureMonitoringStatsAnomaly": { "description": "A list of historical SnapshotAnalysis or ImportFeaturesAnalysis stats requested by user, sorted by FeatureStatsAnomaly.start_time descending.", "id": "GoogleCloudAiplatformV1beta1FeatureMonitoringStatsAnomaly", @@ -27971,6 +28536,36 @@ "properties": {}, "type": "object" }, +"GoogleCloudAiplatformV1beta1FeatureSelectionConfig": { +"description": "Feature selection configuration for the FeatureMonitor.", +"id": "GoogleCloudAiplatformV1beta1FeatureSelectionConfig", +"properties": { +"featureConfigs": { +"description": "Optional. A list of features to be monitored and each feature's drift threshold.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureSelectionConfigFeatureConfig" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1FeatureSelectionConfigFeatureConfig": { +"description": "Feature configuration.", +"id": "GoogleCloudAiplatformV1beta1FeatureSelectionConfigFeatureConfig", +"properties": { +"driftThreshold": { +"description": "Optional. Drift threshold. If calculated difference with baseline data larger than threshold, it will be considered as the feature has drift. If not present, the threshold will be default to 0.3.", +"format": "double", +"type": "number" +}, +"featureId": { +"description": "Required. The ID of the feature resource. Final component of the Feature's resource name.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1FeatureSelector": { "description": "Selector for Features of an EntityType.", "id": "GoogleCloudAiplatformV1beta1FeatureSelector", @@ -27982,6 +28577,49 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1FeatureStatsAndAnomaly": { +"description": "Stats and Anomaly generated by FeatureMonitorJobs. Anomaly only includes Drift.", +"id": "GoogleCloudAiplatformV1beta1FeatureStatsAndAnomaly", +"properties": { +"distributionDeviation": { +"description": "Deviation from the current stats to baseline stats. 1. For categorical feature, the distribution distance is calculated by L-inifinity norm. 2. For numerical feature, the distribution distance is calculated by Jensen\u2013Shannon divergence.", +"format": "double", +"type": "number" +}, +"driftDetected": { +"description": "If set to true, indicates current stats is detected as and comparing with baseline stats.", +"type": "boolean" +}, +"driftDetectionThreshold": { +"description": "This is the threshold used when detecting drifts, which is set in FeatureMonitor.FeatureSelectionConfig.FeatureConfig.drift_threshold", +"format": "double", +"type": "number" +}, +"featureId": { +"description": "Feature Id.", +"type": "string" +}, +"featureMonitorId": { +"description": "The ID of the FeatureMonitor that this FeatureStatsAndAnomaly generated according to.", +"type": "string" +}, +"featureMonitorJobId": { +"description": "The ID of the FeatureMonitorJob that generated this FeatureStatsAndAnomaly.", +"format": "int64", +"type": "string" +}, +"featureStats": { +"description": "Feature stats. e.g. histogram buckets. In the format of tensorflow.metadata.v0.DatasetFeatureStatistics.", +"type": "any" +}, +"statsTime": { +"description": "The timestamp we take snapshot for feature values to generate stats.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1FeatureStatsAnomaly": { "description": "Stats and Anomaly generated at specific timestamp for specific Feature. The start_time and end_time are used to define the time range of the dataset that current stats belongs to, e.g. prediction traffic is bucketed into prediction datasets by time window. If the Dataset is not defined by time window, start_time = end_time. Timestamp of the stats and anomalies always refers to end_time. Raw stats and anomalies are stored in stats_uri or anomaly_uri in the tensorflow defined protos. Field data_stats contains almost identical information with the raw stats in Vertex AI defined proto, for UI to display.", "id": "GoogleCloudAiplatformV1beta1FeatureStatsAnomaly", @@ -28850,6 +29488,17 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1FetchPredictOperationRequest": { +"description": "Request message for PredictionService.FetchPredictOperation.", +"id": "GoogleCloudAiplatformV1beta1FetchPredictOperationRequest", +"properties": { +"operationName": { +"description": "Required. The server-assigned name for the operation.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1FileData": { "description": "URI based data.", "id": "GoogleCloudAiplatformV1beta1FileData", @@ -29233,7 +29882,7 @@ "type": "object" }, "GoogleCloudAiplatformV1beta1FunctionDeclaration": { -"description": "Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client.", +"description": "Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name, description, parameters and response type. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client.", "id": "GoogleCloudAiplatformV1beta1FunctionDeclaration", "properties": { "description": { @@ -31228,6 +31877,42 @@ }, "type": "object" }, +"GoogleCloudAiplatformV1beta1ListFeatureMonitorJobsResponse": { +"description": "Response message for FeatureRegistryService.ListFeatureMonitorJobs.", +"id": "GoogleCloudAiplatformV1beta1ListFeatureMonitorJobsResponse", +"properties": { +"featureMonitorJobs": { +"description": "The FeatureMonitorJobs matching the request.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitorJob" +}, +"type": "array" +}, +"nextPageToken": { +"description": "A token, which can be sent as ListFeatureMonitorJobsRequest.page_token to retrieve the next page. If this field is omitted, there are no subsequent pages.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1ListFeatureMonitorsResponse": { +"description": "Response message for FeatureRegistryService.ListFeatureMonitors.", +"id": "GoogleCloudAiplatformV1beta1ListFeatureMonitorsResponse", +"properties": { +"featureMonitors": { +"description": "The FeatureMonitors matching the request.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1FeatureMonitor" +}, +"type": "array" +}, +"nextPageToken": { +"description": "A token, which can be sent as ListFeatureMonitorsRequest.page_token to retrieve the next page. If this field is omitted, there are no subsequent pages.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1ListFeatureOnlineStoresResponse": { "description": "Response message for FeatureOnlineStoreAdminService.ListFeatureOnlineStores.", "id": "GoogleCloudAiplatformV1beta1ListFeatureOnlineStoresResponse", @@ -35214,6 +35899,10 @@ false "readOnly": true, "type": "string" }, +"customEnvironmentSpec": { +"$ref": "GoogleCloudAiplatformV1beta1NotebookExecutionJobCustomEnvironmentSpec", +"description": "The custom compute configuration for an execution job." +}, "dataformRepositorySource": { "$ref": "GoogleCloudAiplatformV1beta1NotebookExecutionJobDataformRepositorySource", "description": "The Dataform Repository pointing to a single file notebook repository." @@ -35319,6 +36008,25 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1beta1NotebookExecutionJobCustomEnvironmentSpec": { +"description": "Compute configuration to use for an execution job.", +"id": "GoogleCloudAiplatformV1beta1NotebookExecutionJobCustomEnvironmentSpec", +"properties": { +"machineSpec": { +"$ref": "GoogleCloudAiplatformV1beta1MachineSpec", +"description": "The specification of a single machine for the execution job." +}, +"networkSpec": { +"$ref": "GoogleCloudAiplatformV1beta1NetworkSpec", +"description": "The network configuration to use for the execution job." +}, +"persistentDiskSpec": { +"$ref": "GoogleCloudAiplatformV1beta1PersistentDiskSpec", +"description": "The specification of a persistent disk to attach for the execution job." +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1NotebookExecutionJobDataformRepositorySource": { "description": "The Dataform Repository containing the input notebook.", "id": "GoogleCloudAiplatformV1beta1NotebookExecutionJobDataformRepositorySource", @@ -36800,6 +37508,24 @@ false "properties": {}, "type": "object" }, +"GoogleCloudAiplatformV1beta1PredictLongRunningRequest": { +"description": "Request message for PredictionService.PredictLongRunning.", +"id": "GoogleCloudAiplatformV1beta1PredictLongRunningRequest", +"properties": { +"instances": { +"description": "Required. The instances that are the input to the prediction call. A DeployedModel may have an upper limit on the number of instances it supports per request, and when it is exceeded the prediction call errors in case of AutoML Models, or, in case of customer created Models, the behaviour is as documented by that Model. The schema of any single instance may be specified via Endpoint's DeployedModels' Model's PredictSchemata's instance_schema_uri.", +"items": { +"type": "any" +}, +"type": "array" +}, +"parameters": { +"description": "Optional. The parameters that govern the prediction. The schema of the parameters may be specified via Endpoint's DeployedModels' Model's PredictSchemata's parameters_schema_uri.", +"type": "any" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1PredictLongRunningResponse": { "description": "Response message for [PredictionService.PredictLongRunning]", "id": "GoogleCloudAiplatformV1beta1PredictLongRunningResponse", @@ -37687,6 +38413,10 @@ false "description": "Request message for ReasoningEngineExecutionService.Query.", "id": "GoogleCloudAiplatformV1beta1QueryReasoningEngineRequest", "properties": { +"classMethod": { +"description": "Optional. Class method to be used for the query. It is optional and defaults to \"query\" if unspecified.", +"type": "string" +}, "input": { "additionalProperties": { "description": "Properties of the object.", @@ -39862,6 +40592,17 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1beta1ScheduleConfig": { +"description": "Schedule configuration for the FeatureMonitor.", +"id": "GoogleCloudAiplatformV1beta1ScheduleConfig", +"properties": { +"cron": { +"description": "Cron schedule (https://en.wikipedia.org/wiki/Cron) to launch scheduled runs. To explicitly set a timezone to the cron tab, apply a prefix in the cron tab: \"CRON_TZ=${IANA_TIME_ZONE}\" or \"TZ=${IANA_TIME_ZONE}\". The ${IANA_TIME_ZONE} may only be a valid string from IANA time zone database. For example, \"CRON_TZ=America/New_York 1 * * * *\", or \"TZ=America/New_York 1 * * * *\".", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1ScheduleRunResponse": { "description": "Status of a scheduled run.", "id": "GoogleCloudAiplatformV1beta1ScheduleRunResponse", @@ -41883,6 +42624,10 @@ false "structuredPrompt": { "$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecStructuredPrompt", "description": "The prompt variation that stores preambles in separate fields." +}, +"translationPrompt": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationPrompt", +"description": "The prompt variation for Translation use case." } }, "type": "object" @@ -41981,6 +42726,35 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePair": { +"description": "A pair of sentences used as reference in source and target languages.", +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePair", +"properties": { +"sourceSentence": { +"description": "Source sentence in the sentence pair.", +"type": "string" +}, +"targetSentence": { +"description": "Target sentence in the sentence pair.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePairList": { +"description": "A list of reference sentence pairs.", +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePairList", +"properties": { +"referenceSentencePairs": { +"description": "Reference sentence pairs.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePair" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1SchemaPromptSpecStructuredPrompt": { "description": "Prompt variation that stores preambles in separate fields.", "id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecStructuredPrompt", @@ -42032,6 +42806,108 @@ false }, "type": "object" }, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationExample": { +"description": "The translation example that contains reference sentences from various sources.", +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationExample", +"properties": { +"referenceSentencePairLists": { +"description": "The reference sentences from inline text.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecReferenceSentencePairList" +}, +"type": "array" +}, +"referenceSentencesFileInputs": { +"description": "The reference sentences from file.", +"items": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationSentenceFileInput" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationFileInputSource": { +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationFileInputSource", +"properties": { +"content": { +"description": "The file's contents.", +"type": "string" +}, +"displayName": { +"description": "The file's display name.", +"type": "string" +}, +"mimeType": { +"description": "The file's mime type.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationGcsInputSource": { +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationGcsInputSource", +"properties": { +"inputUri": { +"description": "Source data URI. For example, `gs://my_bucket/my_object`.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationOption": { +"description": "Optional settings for translation prompt.", +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationOption", +"properties": { +"numberOfShots": { +"description": "How many shots to use.", +"format": "int32", +"type": "integer" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationPrompt": { +"description": "Prompt variation for Translation use case.", +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationPrompt", +"properties": { +"example": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationExample", +"description": "The translation example." +}, +"option": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationOption", +"description": "The translation option." +}, +"promptMessage": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecPromptMessage", +"description": "The prompt message." +}, +"sourceLanguageCode": { +"description": "The source language code.", +"type": "string" +}, +"targetLanguageCode": { +"description": "The target language code.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationSentenceFileInput": { +"id": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationSentenceFileInput", +"properties": { +"fileInputSource": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationFileInputSource", +"description": "Inlined file source." +}, +"gcsInputSource": { +"$ref": "GoogleCloudAiplatformV1beta1SchemaPromptSpecTranslationGcsInputSource", +"description": "Cloud Storage file source." +} +}, +"type": "object" +}, "GoogleCloudAiplatformV1beta1SchemaTablesDatasetMetadata": { "description": "The metadata of Datasets that contain tables data.", "id": "GoogleCloudAiplatformV1beta1SchemaTablesDatasetMetadata", @@ -47169,7 +48045,7 @@ false "id": "GoogleCloudAiplatformV1beta1TuningJob", "properties": { "baseModel": { -"description": "The base model that is being tuned, e.g., \"gemini-1.0-pro-002\".", +"description": "The base model that is being tuned, e.g., \"gemini-1.0-pro-002\". .", "type": "string" }, "createTime": { diff --git a/googleapiclient/discovery_cache/documents/alloydb.v1.json b/googleapiclient/discovery_cache/documents/alloydb.v1.json index 42cc35d6873..5218276c2c9 100644 --- a/googleapiclient/discovery_cache/documents/alloydb.v1.json +++ b/googleapiclient/discovery_cache/documents/alloydb.v1.json @@ -1517,7 +1517,7 @@ } } }, -"revision": "20241015", +"revision": "20241023", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -2458,7 +2458,7 @@ false "additionalProperties": { "type": "string" }, -"description": "Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", +"description": "Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", "type": "object" }, "deleteTime": { diff --git a/googleapiclient/discovery_cache/documents/alloydb.v1alpha.json b/googleapiclient/discovery_cache/documents/alloydb.v1alpha.json index c06a1121c68..9db9d8864a8 100644 --- a/googleapiclient/discovery_cache/documents/alloydb.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/alloydb.v1alpha.json @@ -1517,7 +1517,7 @@ } } }, -"revision": "20241015", +"revision": "20241023", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -2535,7 +2535,7 @@ false "additionalProperties": { "type": "string" }, -"description": "Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", +"description": "Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", "type": "object" }, "deleteTime": { diff --git a/googleapiclient/discovery_cache/documents/alloydb.v1beta.json b/googleapiclient/discovery_cache/documents/alloydb.v1beta.json index 86031531f8c..41c3d354d41 100644 --- a/googleapiclient/discovery_cache/documents/alloydb.v1beta.json +++ b/googleapiclient/discovery_cache/documents/alloydb.v1beta.json @@ -1514,7 +1514,7 @@ } } }, -"revision": "20241015", +"revision": "20241023", "rootUrl": "https://alloydb.googleapis.com/", "schemas": { "AuthorizedNetwork": { @@ -2516,7 +2516,7 @@ false "additionalProperties": { "type": "string" }, -"description": "Database flags. Set at instance level. * They are copied from primary instance on read instance creation. * Read instances can set new or override existing flags that are relevant for reads, e.g. for enabling columnar cache on a read instance. Flags set on read instance may or may not be present on primary. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", +"description": "Database flags. Set at the instance level. They are copied from the primary instance on secondary instance creation. Flags that have restrictions default to the value at primary instance on read instances during creation. Read instances can set new flags or override existing flags that are relevant for reads, for example, for enabling columnar cache on a read instance. Flags set on read instance might or might not be present on the primary instance. This is a list of \"key\": \"value\" pairs. \"key\": The name of the flag. These flags are passed at instance setup time, so include both server options and system variables for Postgres. Flags are specified with underscores, not hyphens. \"value\": The value of the flag. Booleans are set to **on** for true and **off** for false. This field must be omitted if the flag doesn't take a value.", "type": "object" }, "deleteTime": { diff --git a/googleapiclient/discovery_cache/documents/analyticshub.v1.json b/googleapiclient/discovery_cache/documents/analyticshub.v1.json index ec7088a707a..150b5c5497c 100644 --- a/googleapiclient/discovery_cache/documents/analyticshub.v1.json +++ b/googleapiclient/discovery_cache/documents/analyticshub.v1.json @@ -1022,7 +1022,7 @@ } } }, -"revision": "20240925", +"revision": "20241028", "rootUrl": "https://analyticshub.googleapis.com/", "schemas": { "AnalyticsHubSubscriptionInfo": { @@ -2228,6 +2228,10 @@ "description": "Required. The parent resource path of the Subscription. e.g. `projects/subscriberproject/locations/US`", "type": "string" }, +"destinationDataset": { +"$ref": "DestinationDataset", +"description": "Optional. BigQuery destination dataset to create for the subscriber." +}, "subscriberContact": { "description": "Email of the subscriber.", "type": "string" diff --git a/googleapiclient/discovery_cache/documents/analyticshub.v1beta1.json b/googleapiclient/discovery_cache/documents/analyticshub.v1beta1.json index 5cd9f66a907..3010404eb19 100644 --- a/googleapiclient/discovery_cache/documents/analyticshub.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/analyticshub.v1beta1.json @@ -695,7 +695,7 @@ } } }, -"revision": "20240630", +"revision": "20241028", "rootUrl": "https://analyticshub.googleapis.com/", "schemas": { "AuditConfig": { @@ -833,51 +833,6 @@ }, "type": "object" }, -"DestinationDataset": { -"description": "Defines the destination bigquery dataset.", -"id": "DestinationDataset", -"properties": { -"datasetReference": { -"$ref": "DestinationDatasetReference", -"description": "Required. A reference that identifies the destination dataset." -}, -"description": { -"description": "Optional. A user-friendly description of the dataset.", -"type": "string" -}, -"friendlyName": { -"description": "Optional. A descriptive name for the dataset.", -"type": "string" -}, -"labels": { -"additionalProperties": { -"type": "string" -}, -"description": "Optional. The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See https://cloud.google.com/resource-manager/docs/creating-managing-labels for more information.", -"type": "object" -}, -"location": { -"description": "Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations.", -"type": "string" -} -}, -"type": "object" -}, -"DestinationDatasetReference": { -"description": "Contains the reference that identifies a destination bigquery dataset.", -"id": "DestinationDatasetReference", -"properties": { -"datasetId": { -"description": "Required. A unique ID for this dataset, without the project name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.", -"type": "string" -}, -"projectId": { -"description": "Required. The ID of the project containing this dataset.", -"type": "string" -} -}, -"type": "object" -}, "Empty": { "description": "A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); }", "id": "Empty", @@ -930,6 +885,51 @@ }, "type": "object" }, +"GoogleCloudBigqueryDataexchangeV1beta1DestinationDataset": { +"description": "Defines the destination bigquery dataset.", +"id": "GoogleCloudBigqueryDataexchangeV1beta1DestinationDataset", +"properties": { +"datasetReference": { +"$ref": "GoogleCloudBigqueryDataexchangeV1beta1DestinationDatasetReference", +"description": "Required. A reference that identifies the destination dataset." +}, +"description": { +"description": "Optional. A user-friendly description of the dataset.", +"type": "string" +}, +"friendlyName": { +"description": "Optional. A descriptive name for the dataset.", +"type": "string" +}, +"labels": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. The labels associated with this dataset. You can use these to organize and group your datasets. You can set this property when inserting or updating a dataset. See https://cloud.google.com/resource-manager/docs/creating-managing-labels for more information.", +"type": "object" +}, +"location": { +"description": "Required. The geographic location where the dataset should reside. See https://cloud.google.com/bigquery/docs/locations for supported locations.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudBigqueryDataexchangeV1beta1DestinationDatasetReference": { +"description": "Contains the reference that identifies a destination bigquery dataset.", +"id": "GoogleCloudBigqueryDataexchangeV1beta1DestinationDatasetReference", +"properties": { +"datasetId": { +"description": "Required. A unique ID for this dataset, without the project name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.", +"type": "string" +}, +"projectId": { +"description": "Required. The ID of the project containing this dataset.", +"type": "string" +} +}, +"type": "object" +}, "LinkedResource": { "description": "Reference to a linked resource tracked by this Subscription.", "id": "LinkedResource", @@ -1276,7 +1276,7 @@ "id": "SubscribeListingRequest", "properties": { "destinationDataset": { -"$ref": "DestinationDataset", +"$ref": "GoogleCloudBigqueryDataexchangeV1beta1DestinationDataset", "description": "BigQuery destination dataset to create for the subscriber." } }, diff --git a/googleapiclient/discovery_cache/documents/androidmanagement.v1.json b/googleapiclient/discovery_cache/documents/androidmanagement.v1.json index 43e6d9a6e47..5b8c5b537b3 100644 --- a/googleapiclient/discovery_cache/documents/androidmanagement.v1.json +++ b/googleapiclient/discovery_cache/documents/androidmanagement.v1.json @@ -1168,7 +1168,7 @@ } } }, -"revision": "20241022", +"revision": "20241031", "rootUrl": "https://androidmanagement.googleapis.com/", "schemas": { "AdbShellCommandEvent": { @@ -3216,7 +3216,7 @@ false "type": "object" }, "ExtensionConfig": { -"description": "Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 13 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 13 and above.", +"description": "Configuration to enable an app as an extension app, with the capability of interacting with Android Device Policy offline. For Android versions 11 and above, extension apps are exempt from battery restrictions so will not be placed into the restricted App Standby Bucket (https://developer.android.com/topic/performance/appstandby#restricted-bucket). Extensions apps are also protected against users clearing their data or force-closing the application, although admins can continue to use the clear app data command on extension apps if needed for Android 11 and above.", "id": "ExtensionConfig", "properties": { "notificationReceiver": { diff --git a/googleapiclient/discovery_cache/documents/backupdr.v1.json b/googleapiclient/discovery_cache/documents/backupdr.v1.json index 7b002f9c3ed..6255a9a62c3 100644 --- a/googleapiclient/discovery_cache/documents/backupdr.v1.json +++ b/googleapiclient/discovery_cache/documents/backupdr.v1.json @@ -1663,7 +1663,7 @@ } } }, -"revision": "20241017", +"revision": "20241023", "rootUrl": "https://backupdr.googleapis.com/", "schemas": { "AbandonBackupRequest": { @@ -3698,7 +3698,7 @@ "type": "string" }, "networks": { -"description": "Required. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported.", +"description": "Optional. VPC networks to which the ManagementServer instance is connected. For this version, only a single network is supported. This field is optional if MS is created without PSA", "items": { "$ref": "NetworkConfig" }, diff --git a/googleapiclient/discovery_cache/documents/beyondcorp.v1.json b/googleapiclient/discovery_cache/documents/beyondcorp.v1.json index 92fc3438415..63d12db4129 100644 --- a/googleapiclient/discovery_cache/documents/beyondcorp.v1.json +++ b/googleapiclient/discovery_cache/documents/beyondcorp.v1.json @@ -2169,6 +2169,34 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, +"setIamPolicy": { +"description": "Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/securityGateways/{securityGatewaysId}:setIamPolicy", +"httpMethod": "POST", +"id": "beyondcorp.projects.locations.securityGateways.setIamPolicy", +"parameterOrder": [ +"resource" +], +"parameters": { +"resource": { +"description": "REQUIRED: The resource for which the policy is being specified. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/securityGateways/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1/{+resource}:setIamPolicy", +"request": { +"$ref": "GoogleIamV1SetIamPolicyRequest" +}, +"response": { +"$ref": "GoogleIamV1Policy" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, "setPeering": { "description": "This is a custom method to allow customers to create a peering connections between Google network and customer networks. This is enabled only for the allowlisted customers.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/securityGateways/{securityGatewaysId}:setPeering", @@ -2401,53 +2429,9 @@ } } } -}, -"v": { -"resources": { -"projects": { -"resources": { -"locations": { -"resources": { -"securityGateways": { -"methods": { -"setIamPolicy": { -"description": "Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.", -"flatPath": "v/projects/{projectsId}/locations/{locationsId}/securityGateways/{securityGatewaysId}:setIamPolicy", -"httpMethod": "POST", -"id": "beyondcorp.v.projects.locations.securityGateways.setIamPolicy", -"parameterOrder": [ -"resource" -], -"parameters": { -"resource": { -"description": "REQUIRED: The resource for which the policy is being specified. See [Resource names](https://cloud.google.com/apis/design/resource_names) for the appropriate value for this field.", -"location": "path", -"pattern": "^projects/[^/]+/locations/[^/]+/securityGateways/[^/]+$", -"required": true, -"type": "string" -} -}, -"path": "v/{+resource}:setIamPolicy", -"request": { -"$ref": "GoogleIamV1SetIamPolicyRequest" -}, -"response": { -"$ref": "GoogleIamV1Policy" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] -} -} -} -} -} -} -} -} } }, -"revision": "20241011", +"revision": "20241024", "rootUrl": "https://beyondcorp.googleapis.com/", "schemas": { "AllocatedConnection": { diff --git a/googleapiclient/discovery_cache/documents/chat.v1.json b/googleapiclient/discovery_cache/documents/chat.v1.json index c0a2798dca5..f970601823c 100644 --- a/googleapiclient/discovery_cache/documents/chat.v1.json +++ b/googleapiclient/discovery_cache/documents/chat.v1.json @@ -567,7 +567,7 @@ "type": "string" }, "useAdminAccess": { -"description": "When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Creating app memberships or creating memberships for users outside the administrator's Google Workspace organization isn't supported using admin access.", +"description": "Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Creating app memberships or creating memberships for users outside the administrator's Google Workspace organization isn't supported using admin access.", "location": "query", "type": "boolean" } @@ -604,7 +604,7 @@ "type": "string" }, "useAdminAccess": { -"description": "When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Deleting app memberships in a space isn't supported using admin access.", +"description": "Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Deleting app memberships in a space isn't supported using admin access.", "location": "query", "type": "boolean" } @@ -638,7 +638,7 @@ "type": "string" }, "useAdminAccess": { -"description": "When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` or `chat.admin.memberships.readonly` [OAuth 2.0 scopes](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Getting app memberships in a space isn't supported when using admin access.", +"description": "Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` or `chat.admin.memberships.readonly` [OAuth 2.0 scopes](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Getting app memberships in a space isn't supported when using admin access.", "location": "query", "type": "boolean" } @@ -698,7 +698,7 @@ "type": "boolean" }, "useAdminAccess": { -"description": "When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires either the `chat.admin.memberships.readonly` or `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Listing app memberships in a space isn't supported when using admin access.", +"description": "Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires either the `chat.admin.memberships.readonly` or `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes). Listing app memberships in a space isn't supported when using admin access.", "location": "query", "type": "boolean" } @@ -726,7 +726,7 @@ ], "parameters": { "name": { -"description": "Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`", +"description": "Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`", "location": "path", "pattern": "^spaces/[^/]+/members/[^/]+$", "required": true, @@ -739,7 +739,7 @@ "type": "string" }, "useAdminAccess": { -"description": "When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes).", +"description": "Optional. When `true`, the method runs using the user's Google Workspace administrator privileges. The calling user must be a Google Workspace administrator with the [manage chat and spaces conversations privilege](https://support.google.com/a/answer/13369245). Requires the `chat.admin.memberships` [OAuth 2.0 scope](https://developers.google.com/workspace/chat/authenticate-authorize#chat-api-scopes).", "location": "query", "type": "boolean" } @@ -1212,7 +1212,7 @@ "type": "integer" }, "pageToken": { -"description": "A page token, received from a previous list space events call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to list space events must match the call that provided the page token. Passing different values to the other parameters might lead to unexpected results.", +"description": "Optional. A page token, received from a previous list space events call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to list space events must match the call that provided the page token. Passing different values to the other parameters might lead to unexpected results.", "location": "query", "type": "string" }, @@ -1344,7 +1344,7 @@ } } }, -"revision": "20241020", +"revision": "20241029", "rootUrl": "https://chat.googleapis.com/", "schemas": { "AccessSettings": { @@ -3774,14 +3774,14 @@ }, "groupMember": { "$ref": "Group", -"description": "The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user)." +"description": "Optional. The Google Group the membership corresponds to. Reading or mutating memberships for Google Groups requires [user authentication](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user)." }, "member": { "$ref": "User", -"description": "The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`." +"description": "Optional. The Google Chat user or app the membership corresponds to. If your Chat app [authenticates as a user](https://developers.google.com/workspace/chat/authenticate-authorize-chat-user), the output populates the [user](https://developers.google.com/workspace/chat/api/reference/rest/v1/User) `name` and `type`." }, "name": { -"description": "Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`", +"description": "Identifier. Resource name of the membership, assigned by the server. Format: `spaces/{space}/members/{member}`", "type": "string" }, "role": { diff --git a/googleapiclient/discovery_cache/documents/clouddeploy.v1.json b/googleapiclient/discovery_cache/documents/clouddeploy.v1.json index 97f0a66ea17..a9b562beee2 100644 --- a/googleapiclient/discovery_cache/documents/clouddeploy.v1.json +++ b/googleapiclient/discovery_cache/documents/clouddeploy.v1.json @@ -2289,7 +2289,7 @@ } } }, -"revision": "20241010", +"revision": "20241029", "rootUrl": "https://clouddeploy.googleapis.com/", "schemas": { "AbandonReleaseRequest": { @@ -2694,6 +2694,10 @@ true "repairRolloutRule": { "$ref": "RepairRolloutRule", "description": "Optional. The `RepairRolloutRule` will automatically repair a failed rollout." +}, +"timedPromoteReleaseRule": { +"$ref": "TimedPromoteReleaseRule", +"description": "Optional. The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule." } }, "type": "object" @@ -2705,6 +2709,10 @@ true "targetsPresentCondition": { "$ref": "TargetsPresentCondition", "description": "Optional. Details around targets enumerated in the rule." +}, +"timedPromoteReleaseCondition": { +"$ref": "TimedPromoteReleaseCondition", +"description": "Optional. TimedPromoteReleaseCondition contains rule conditions specific to a an Automation with a timed promote release rule defined." } }, "type": "object" @@ -2804,10 +2812,15 @@ true "type": "string" }, "targetId": { -"description": "Output only. The ID of the target that represents the promotion stage that initiates the `AutomationRun`. The value of this field is the last segment of a target name.", +"description": "Output only. The ID of the source target that initiates the `AutomationRun`. The value of this field is the last segment of a target name.", "readOnly": true, "type": "string" }, +"timedPromoteReleaseOperation": { +"$ref": "TimedPromoteReleaseOperation", +"description": "Output only. Promotes a release to a specified 'Target' as defined in a Timed Promote Release rule.", +"readOnly": true +}, "updateTime": { "description": "Output only. Time at which the automationRun was updated.", "format": "google-datetime", @@ -6914,6 +6927,21 @@ true }, "type": "object" }, +"Targets": { +"description": "The targets involved in a single timed promotion.", +"id": "Targets", +"properties": { +"destinationTargetId": { +"description": "Optional. The destination target ID.", +"type": "string" +}, +"sourceTargetId": { +"description": "Optional. The source target ID.", +"type": "string" +} +}, +"type": "object" +}, "TargetsPresentCondition": { "description": "`TargetsPresentCondition` contains information on any Targets referenced in the Delivery Pipeline that do not actually exist.", "id": "TargetsPresentCondition", @@ -7052,6 +7080,81 @@ true }, "type": "object" }, +"TimedPromoteReleaseCondition": { +"description": "`TimedPromoteReleaseCondition` contains conditions specific to an Automation with a Timed Promote Release rule defined.", +"id": "TimedPromoteReleaseCondition", +"properties": { +"nextPromotionTime": { +"description": "Output only. When the next scheduled promotion(s) will occur.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"targetsList": { +"description": "Output only. A list of targets involved in the upcoming timed promotion(s).", +"items": { +"$ref": "Targets" +}, +"readOnly": true, +"type": "array" +} +}, +"type": "object" +}, +"TimedPromoteReleaseOperation": { +"description": "Contains the information of an automated timed promote-release operation.", +"id": "TimedPromoteReleaseOperation", +"properties": { +"phase": { +"description": "Output only. The starting phase of the rollout created by this operation.", +"readOnly": true, +"type": "string" +}, +"release": { +"description": "Output only. The name of the release to be promoted.", +"readOnly": true, +"type": "string" +}, +"targetId": { +"description": "Output only. The ID of the target that represents the promotion stage to which the release will be promoted. The value of this field is the last segment of a target name.", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"TimedPromoteReleaseRule": { +"description": "The `TimedPromoteReleaseRule` will automatically promote a release from the current target(s) to the specified target(s) on a configured schedule.", +"id": "TimedPromoteReleaseRule", +"properties": { +"condition": { +"$ref": "AutomationRuleCondition", +"description": "Output only. Information around the state of the Automation rule.", +"readOnly": true +}, +"destinationPhase": { +"description": "Optional. The starting phase of the rollout created by this rule. Default to the first phase.", +"type": "string" +}, +"destinationTargetId": { +"description": "Optional. The ID of the stage in the pipeline to which this `Release` is deploying. If unspecified, default it to the next stage in the promotion flow. The value of this field could be one of the following: * The last segment of a target name * \"@next\", the next target in the promotion sequence", +"type": "string" +}, +"id": { +"description": "Required. ID of the rule. This ID must be unique in the `Automation` resource to which this rule belongs. The format is `[a-z]([a-z0-9-]{0,61}[a-z0-9])?`.", +"type": "string" +}, +"schedule": { +"description": "Required. Schedule in crontab format. e.g. \"0 9 * * 1\" for every Monday at 9am.", +"type": "string" +}, +"timeZone": { +"description": "Required. The time zone in IANA format [IANA Time Zone Database](https://www.iana.org/time-zones) (e.g. America/New_York).", +"type": "string" +} +}, +"type": "object" +}, "VerifyJob": { "description": "A verify Job.", "id": "VerifyJob", diff --git a/googleapiclient/discovery_cache/documents/cloudfunctions.v2.json b/googleapiclient/discovery_cache/documents/cloudfunctions.v2.json index ff730966bf5..ff925af736d 100644 --- a/googleapiclient/discovery_cache/documents/cloudfunctions.v2.json +++ b/googleapiclient/discovery_cache/documents/cloudfunctions.v2.json @@ -716,7 +716,7 @@ } } }, -"revision": "20240905", +"revision": "20241024", "rootUrl": "https://cloudfunctions.googleapis.com/", "schemas": { "AbortFunctionUpgradeRequest": { @@ -2058,7 +2058,7 @@ "type": "object" }, "UpgradeInfo": { -"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration.", +"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach.", "id": "UpgradeInfo", "properties": { "buildConfig": { @@ -2085,7 +2085,8 @@ "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_SUCCESSFUL", "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_ERROR", "ROLLBACK_FUNCTION_UPGRADE_TRAFFIC_ERROR", -"COMMIT_FUNCTION_UPGRADE_ERROR" +"COMMIT_FUNCTION_UPGRADE_ERROR", +"DETACH_IN_PROGRESS" ], "enumDescriptions": [ "Unspecified state. Most functions are in this upgrade state.", @@ -2097,7 +2098,8 @@ "RedirectFunctionUpgradeTraffic API was successful and traffic is served by 2nd Gen function stack.", "RedirectFunctionUpgradeTraffic API was un-successful.", "RollbackFunctionUpgradeTraffic API was un-successful.", -"CommitFunctionUpgrade API was un-successful." +"CommitFunctionUpgrade API was un-successful.", +"Function is requested to be detached from 2nd Gen to CRf." ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/cloudfunctions.v2alpha.json b/googleapiclient/discovery_cache/documents/cloudfunctions.v2alpha.json index 4db5bdb2346..5101c73f705 100644 --- a/googleapiclient/discovery_cache/documents/cloudfunctions.v2alpha.json +++ b/googleapiclient/discovery_cache/documents/cloudfunctions.v2alpha.json @@ -716,7 +716,7 @@ } } }, -"revision": "20240905", +"revision": "20241024", "rootUrl": "https://cloudfunctions.googleapis.com/", "schemas": { "AbortFunctionUpgradeRequest": { @@ -2058,7 +2058,7 @@ "type": "object" }, "UpgradeInfo": { -"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration.", +"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach.", "id": "UpgradeInfo", "properties": { "buildConfig": { @@ -2085,7 +2085,8 @@ "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_SUCCESSFUL", "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_ERROR", "ROLLBACK_FUNCTION_UPGRADE_TRAFFIC_ERROR", -"COMMIT_FUNCTION_UPGRADE_ERROR" +"COMMIT_FUNCTION_UPGRADE_ERROR", +"DETACH_IN_PROGRESS" ], "enumDescriptions": [ "Unspecified state. Most functions are in this upgrade state.", @@ -2097,7 +2098,8 @@ "RedirectFunctionUpgradeTraffic API was successful and traffic is served by 2nd Gen function stack.", "RedirectFunctionUpgradeTraffic API was un-successful.", "RollbackFunctionUpgradeTraffic API was un-successful.", -"CommitFunctionUpgrade API was un-successful." +"CommitFunctionUpgrade API was un-successful.", +"Function is requested to be detached from 2nd Gen to CRf." ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/cloudfunctions.v2beta.json b/googleapiclient/discovery_cache/documents/cloudfunctions.v2beta.json index 1bb7552541c..6b1d1f89b2b 100644 --- a/googleapiclient/discovery_cache/documents/cloudfunctions.v2beta.json +++ b/googleapiclient/discovery_cache/documents/cloudfunctions.v2beta.json @@ -716,7 +716,7 @@ } } }, -"revision": "20240905", +"revision": "20241024", "rootUrl": "https://cloudfunctions.googleapis.com/", "schemas": { "AbortFunctionUpgradeRequest": { @@ -2058,7 +2058,7 @@ "type": "object" }, "UpgradeInfo": { -"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration * Current state of migration for function undergoing migration.", +"description": "Information related to: * A function's eligibility for 1st Gen to 2nd Gen migration and 2nd Gen to CRf detach. * Current state of migration for function undergoing migration/detach.", "id": "UpgradeInfo", "properties": { "buildConfig": { @@ -2085,7 +2085,8 @@ "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_SUCCESSFUL", "REDIRECT_FUNCTION_UPGRADE_TRAFFIC_ERROR", "ROLLBACK_FUNCTION_UPGRADE_TRAFFIC_ERROR", -"COMMIT_FUNCTION_UPGRADE_ERROR" +"COMMIT_FUNCTION_UPGRADE_ERROR", +"DETACH_IN_PROGRESS" ], "enumDescriptions": [ "Unspecified state. Most functions are in this upgrade state.", @@ -2097,7 +2098,8 @@ "RedirectFunctionUpgradeTraffic API was successful and traffic is served by 2nd Gen function stack.", "RedirectFunctionUpgradeTraffic API was un-successful.", "RollbackFunctionUpgradeTraffic API was un-successful.", -"CommitFunctionUpgrade API was un-successful." +"CommitFunctionUpgrade API was un-successful.", +"Function is requested to be detached from 2nd Gen to CRf." ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/composer.v1.json b/googleapiclient/discovery_cache/documents/composer.v1.json index 2bdaf12ad63..f18db53d23b 100644 --- a/googleapiclient/discovery_cache/documents/composer.v1.json +++ b/googleapiclient/discovery_cache/documents/composer.v1.json @@ -1188,7 +1188,7 @@ } } }, -"revision": "20241013", +"revision": "20241027", "rootUrl": "https://composer.googleapis.com/", "schemas": { "AirflowMetadataRetentionPolicyConfig": { @@ -1508,7 +1508,7 @@ "properties": { "config": { "$ref": "EnvironmentConfig", -"description": "Configuration parameters for this environment." +"description": "Optional. Configuration parameters for this environment." }, "createTime": { "description": "Output only. The time at which this environment was created.", @@ -1523,7 +1523,7 @@ "type": "object" }, "name": { -"description": "The resource name of the environment, in the form: \"projects/{projectId}/locations/{locationId}/environments/{environmentId}\" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen.", +"description": "Identifier. The resource name of the environment, in the form: \"projects/{projectId}/locations/{locationId}/environments/{environmentId}\" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen.", "type": "string" }, "satisfiesPzi": { @@ -1631,7 +1631,7 @@ }, "nodeConfig": { "$ref": "NodeConfig", -"description": "The configuration used for the Kubernetes Engine cluster." +"description": "Optional. The configuration used for the Kubernetes Engine cluster." }, "nodeCount": { "description": "The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*.", @@ -1640,7 +1640,7 @@ }, "privateEnvironmentConfig": { "$ref": "PrivateEnvironmentConfig", -"description": "The configuration used for the Private IP Cloud Composer environment." +"description": "Optional. The configuration used for the Private IP Cloud Composer environment." }, "recoveryConfig": { "$ref": "RecoveryConfig", @@ -1660,7 +1660,7 @@ }, "softwareConfig": { "$ref": "SoftwareConfig", -"description": "The configuration settings for software inside the environment." +"description": "Optional. The configuration settings for software inside the environment." }, "webServerConfig": { "$ref": "WebServerConfig", @@ -2426,7 +2426,7 @@ "type": "object" }, "imageVersion": { -"description": "The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\\.[0-9]+\\.[0-9]+(-preview\\.[0-9]+)?)?|latest)-airflow-([0-9]+(\\.[0-9]+(\\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview).", +"description": "Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\\.[0-9]+\\.[0-9]+(-preview\\.[0-9]+)?)?|latest)-airflow-([0-9]+(\\.[0-9]+(\\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview).", "type": "string" }, "pypiPackages": { @@ -2592,7 +2592,7 @@ "additionalProperties": { "type": "string" }, -"description": "Optional. The \"data\" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/", +"description": "Optional. The \"data\" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { \"example_key\": \"example_value\", \"another_key\": \"another_value\" }", "type": "object" }, "name": { @@ -2610,7 +2610,7 @@ "additionalProperties": { "type": "string" }, -"description": "Optional. The \"data\" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/", +"description": "Optional. The \"data\" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { \"example\": \"ZXhhbXBsZV92YWx1ZQ==\", \"another-example\": \"YW5vdGhlcl9leGFtcGxlX3ZhbHVl\" }", "type": "object" }, "name": { diff --git a/googleapiclient/discovery_cache/documents/composer.v1beta1.json b/googleapiclient/discovery_cache/documents/composer.v1beta1.json index fa7338aa3f2..bad0a43d2ab 100644 --- a/googleapiclient/discovery_cache/documents/composer.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/composer.v1beta1.json @@ -1216,7 +1216,7 @@ } } }, -"revision": "20241013", +"revision": "20241027", "rootUrl": "https://composer.googleapis.com/", "schemas": { "AirflowMetadataRetentionPolicyConfig": { @@ -1542,7 +1542,7 @@ "properties": { "config": { "$ref": "EnvironmentConfig", -"description": "Configuration parameters for this environment." +"description": "Optional. Configuration parameters for this environment." }, "createTime": { "description": "Output only. The time at which this environment was created.", @@ -1558,7 +1558,7 @@ "type": "object" }, "name": { -"description": "The resource name of the environment, in the form: \"projects/{projectId}/locations/{locationId}/environments/{environmentId}\" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen.", +"description": "Identifier. The resource name of the environment, in the form: \"projects/{projectId}/locations/{locationId}/environments/{environmentId}\" EnvironmentId must start with a lowercase letter followed by up to 63 lowercase letters, numbers, or hyphens, and cannot end with a hyphen.", "type": "string" }, "satisfiesPzi": { @@ -1671,7 +1671,7 @@ }, "nodeConfig": { "$ref": "NodeConfig", -"description": "The configuration used for the Kubernetes Engine cluster." +"description": "Optional. The configuration used for the Kubernetes Engine cluster." }, "nodeCount": { "description": "The number of nodes in the Kubernetes Engine cluster that will be used to run this environment. This field is supported for Cloud Composer environments in versions composer-1.*.*-airflow-*.*.*.", @@ -1680,7 +1680,7 @@ }, "privateEnvironmentConfig": { "$ref": "PrivateEnvironmentConfig", -"description": "The configuration used for the Private IP Cloud Composer environment." +"description": "Optional. The configuration used for the Private IP Cloud Composer environment." }, "recoveryConfig": { "$ref": "RecoveryConfig", @@ -1700,7 +1700,7 @@ }, "softwareConfig": { "$ref": "SoftwareConfig", -"description": "The configuration settings for software inside the environment." +"description": "Optional. The configuration settings for software inside the environment." }, "webServerConfig": { "$ref": "WebServerConfig", @@ -2475,7 +2475,7 @@ "type": "object" }, "imageVersion": { -"description": "The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\\.[0-9]+\\.[0-9]+(-preview\\.[0-9]+)?)?|latest)-airflow-([0-9]+(\\.[0-9]+(\\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview).", +"description": "Optional. The version of the software running in the environment. This encapsulates both the version of Cloud Composer functionality and the version of Apache Airflow. It must match the regular expression `composer-([0-9]+(\\.[0-9]+\\.[0-9]+(-preview\\.[0-9]+)?)?|latest)-airflow-([0-9]+(\\.[0-9]+(\\.[0-9]+)?)?)`. When used as input, the server also checks if the provided version is supported and denies the request for an unsupported version. The Cloud Composer portion of the image version is a full [semantic version](https://semver.org), or an alias in the form of major version number or `latest`. When an alias is provided, the server replaces it with the current Cloud Composer version that satisfies the alias. The Apache Airflow portion of the image version is a full semantic version that points to one of the supported Apache Airflow versions, or an alias in the form of only major or major.minor versions specified. When an alias is provided, the server replaces it with the latest Apache Airflow version that satisfies the alias and is supported in the given Cloud Composer version. In all cases, the resolved image version is stored in the same field. See also [version list](/composer/docs/concepts/versioning/composer-versions) and [versioning overview](/composer/docs/concepts/versioning/composer-versioning-overview).", "type": "string" }, "pypiPackages": { @@ -2641,7 +2641,7 @@ "additionalProperties": { "type": "string" }, -"description": "Optional. The \"data\" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/", +"description": "Optional. The \"data\" field of Kubernetes ConfigMap, organized in key-value pairs. For details see: https://kubernetes.io/docs/concepts/configuration/configmap/ Example: { \"example_key\": \"example_value\", \"another_key\": \"another_value\" }", "type": "object" }, "name": { @@ -2659,7 +2659,7 @@ "additionalProperties": { "type": "string" }, -"description": "Optional. The \"data\" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/", +"description": "Optional. The \"data\" field of Kubernetes Secret, organized in key-value pairs, which can contain sensitive values such as a password, a token, or a key. The values for all keys have to be base64-encoded strings. For details see: https://kubernetes.io/docs/concepts/configuration/secret/ Example: { \"example\": \"ZXhhbXBsZV92YWx1ZQ==\", \"another-example\": \"YW5vdGhlcl9leGFtcGxlX3ZhbHVl\" }", "type": "object" }, "name": { diff --git a/googleapiclient/discovery_cache/documents/compute.v1.json b/googleapiclient/discovery_cache/documents/compute.v1.json index 6ba9d922b9e..2cb40c20260 100644 --- a/googleapiclient/discovery_cache/documents/compute.v1.json +++ b/googleapiclient/discovery_cache/documents/compute.v1.json @@ -10446,7 +10446,7 @@ "type": "string" }, "start": { -"description": "Specifies the starting byte position of the output to return. To start with the first byte of output to the specified port, omit this field or set it to `0`. If the output for that byte position is available, this field matches the `start` parameter sent with the request. If the amount of serial console output exceeds the size of the buffer (1 MB), the oldest output is discarded and is no longer available. If the requested start position refers to discarded output, the start position is adjusted to the oldest output still available, and the adjusted start position is returned as the `start` property value. You can also provide a negative start position, which translates to the most recent number of bytes written to the serial port. For example, -3 is interpreted as the most recent 3 bytes written to the serial console.", +"description": "Specifies the starting byte position of the output to return. To start with the first byte of output to the specified port, omit this field or set it to `0`. If the output for that byte position is available, this field matches the `start` parameter sent with the request. If the amount of serial console output exceeds the size of the buffer (1 MB), the oldest output is discarded and is no longer available. If the requested start position refers to discarded output, the start position is adjusted to the oldest output still available, and the adjusted start position is returned as the `start` property value. You can also provide a negative start position, which translates to the most recent number of bytes written to the serial port. For example, -3 is interpreted as the most recent 3 bytes written to the serial console. Note that the negative start is bounded by the retained buffer size, and the returned serial console output will not exceed the max buffer size.", "format": "int64", "location": "query", "type": "string" @@ -37421,7 +37421,7 @@ } } }, -"revision": "20241015", +"revision": "20241021", "rootUrl": "https://compute.googleapis.com/", "schemas": { "AWSV4Signature": { @@ -40146,7 +40146,7 @@ false "description": "The CacheKeyPolicy for this CdnPolicy." }, "cacheMode": { -"description": "Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any \"private\", \"no-store\" or \"no-cache\" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached.", +"description": "Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any \"private\", \"no-store\" or \"no-cache\" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC.", "enum": [ "CACHE_ALL_STATIC", "FORCE_CACHE_ALL", @@ -40908,7 +40908,7 @@ false "description": "The CacheKeyPolicy for this CdnPolicy." }, "cacheMode": { -"description": "Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any \"private\", \"no-store\" or \"no-cache\" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached.", +"description": "Specifies the cache setting for all responses from this backend. The possible values are: USE_ORIGIN_HEADERS Requires the origin to set valid caching headers to cache content. Responses without these headers will not be cached at Google's edge, and will require a full trip to the origin on every request, potentially impacting performance and increasing load on the origin server. FORCE_CACHE_ALL Cache all content, ignoring any \"private\", \"no-store\" or \"no-cache\" directives in Cache-Control response headers. Warning: this may result in Cloud CDN caching private, per-user (user identifiable) content. CACHE_ALL_STATIC Automatically cache static content, including common image formats, media (video and audio), and web assets (JavaScript and CSS). Requests and responses that are marked as uncacheable, as well as dynamic content (including HTML), will not be cached. If no value is provided for cdnPolicy.cacheMode, it defaults to CACHE_ALL_STATIC.", "enum": [ "CACHE_ALL_STATIC", "FORCE_CACHE_ALL", @@ -60238,6 +60238,34 @@ false "description": "A routing configuration attached to a network resource. The message includes the list of routers associated with the network, and a flag indicating the type of routing behavior to enforce network-wide.", "id": "NetworkRoutingConfig", "properties": { +"bgpAlwaysCompareMed": { +"description": "Enable comparison of Multi-Exit Discriminators (MED) across routes with different neighbor ASNs when using the STANDARD BGP best path selection algorithm.", +"type": "boolean" +}, +"bgpBestPathSelectionMode": { +"description": "The BGP best path selection algorithm to be employed within this network for dynamic routes learned by Cloud Routers. Can be LEGACY (default) or STANDARD.", +"enum": [ +"LEGACY", +"STANDARD" +], +"enumDescriptions": [ +"", +"" +], +"type": "string" +}, +"bgpInterRegionCost": { +"description": "Allows to define a preferred approach for handling inter-region cost in the selection process when using the STANDARD BGP best path selection algorithm. Can be DEFAULT or ADD_COST_TO_MED.", +"enum": [ +"ADD_COST_TO_MED", +"DEFAULT" +], +"enumDescriptions": [ +"", +"" +], +"type": "string" +}, "routingMode": { "description": "The network-wide routing mode to use. If set to REGIONAL, this network's Cloud Routers will only advertise routes with subnets of this network in the same region as the router. If set to GLOBAL, this network's Cloud Routers will advertise routes with all subnets of this network, across regions.", "enum": [ @@ -69127,14 +69155,38 @@ false "description": "The URL to an instance that should handle matching packets. You can specify this as a full or partial URL. For example: https://www.googleapis.com/compute/v1/projects/project/zones/zone/instances/", "type": "string" }, +"nextHopInterRegionCost": { +"description": "[Output only] Internal fixed region-to-region cost that Google Cloud calculates based on factors such as network performance, distance, and available bandwidth between regions.", +"format": "uint32", +"type": "integer" +}, "nextHopIp": { "description": "The network IP address of an instance that should handle matching packets. Both IPv6 address and IPv4 addresses are supported. Must specify an IPv4 address in dot-decimal notation (e.g. 192.0.2.99) or an IPv6 address in RFC 4291 format (e.g. 2001:db8::2d9:51:0:0 or 2001:db8:0:0:2d9:51:0:0). IPv6 addresses will be displayed using RFC 5952 compressed format (e.g. 2001:db8::2d9:51:0:0). Should never be an IPv4-mapped IPv6 address.", "type": "string" }, +"nextHopMed": { +"description": "[Output Only] Multi-Exit Discriminator, a BGP route metric that indicates the desirability of a particular route in a network.", +"format": "uint32", +"type": "integer" +}, "nextHopNetwork": { "description": "The URL of the local network if it should handle matching packets.", "type": "string" }, +"nextHopOrigin": { +"description": "[Output Only] Indicates the origin of the route. Can be IGP (Interior Gateway Protocol), EGP (Exterior Gateway Protocol), or INCOMPLETE.", +"enum": [ +"EGP", +"IGP", +"INCOMPLETE" +], +"enumDescriptions": [ +"", +"", +"" +], +"type": "string" +}, "nextHopPeering": { "description": "[Output Only] The network peering name that should handle matching packets, which should conform to RFC1035.", "type": "string" @@ -76920,7 +76972,7 @@ false "type": "string" }, "purpose": { -"description": "The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY.", +"description": "The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY.", "enum": [ "GLOBAL_MANAGED_PROXY", "INTERNAL_HTTPS_LOAD_BALANCER", @@ -82115,7 +82167,7 @@ false "type": "string" }, "purpose": { -"description": "The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, PRIVATE_SERVICE_CONNECT, or PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY.", +"description": "The purpose of the resource. This field can be either PRIVATE, GLOBAL_MANAGED_PROXY, REGIONAL_MANAGED_PROXY, or PRIVATE_SERVICE_CONNECT. PRIVATE is the default purpose for user-created subnets or subnets that are automatically created in auto mode networks. Subnets with purpose set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY are user-created subnetworks that are reserved for Envoy-based load balancers. A subnet with purpose set to PRIVATE_SERVICE_CONNECT is used to publish services using Private Service Connect. If unspecified, the subnet purpose defaults to PRIVATE. The enableFlowLogs field isn't supported if the subnet purpose field is set to GLOBAL_MANAGED_PROXY or REGIONAL_MANAGED_PROXY.", "enum": [ "GLOBAL_MANAGED_PROXY", "INTERNAL_HTTPS_LOAD_BALANCER", diff --git a/googleapiclient/discovery_cache/documents/container.v1beta1.json b/googleapiclient/discovery_cache/documents/container.v1beta1.json index 2ff7a09beae..6182459150d 100644 --- a/googleapiclient/discovery_cache/documents/container.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/container.v1beta1.json @@ -2685,7 +2685,7 @@ } } }, -"revision": "20241008", +"revision": "20241017", "rootUrl": "https://container.googleapis.com/", "schemas": { "AcceleratorConfig": { @@ -3641,6 +3641,10 @@ "$ref": "ParentProductConfig", "description": "The configuration of the parent product of the cluster. This field is used by Google internal products that are built on top of the GKE cluster and take the ownership of the cluster." }, +"podAutoscaling": { +"$ref": "PodAutoscaling", +"description": "The config for pod autoscaling." +}, "podSecurityPolicyConfig": { "$ref": "PodSecurityPolicyConfig", "description": "Configuration for the PodSecurityPolicy feature." @@ -3961,6 +3965,10 @@ "description": "Enable/Disable private endpoint for the cluster's master. Deprecated: Use desired_control_plane_endpoints_config.ip_endpoints_config.enable_public_endpoint instead. Note that the value of enable_public_endpoint is reversed: if enable_private_endpoint is false, then enable_public_endpoint will be true.", "type": "boolean" }, +"desiredEnterpriseConfig": { +"$ref": "DesiredEnterpriseConfig", +"description": "The desired enterprise configuration for the cluster." +}, "desiredFleet": { "$ref": "Fleet", "description": "The desired fleet configuration for the cluster." @@ -4095,6 +4103,10 @@ "$ref": "ParentProductConfig", "description": "The desired parent product config for the cluster." }, +"desiredPodAutoscaling": { +"$ref": "PodAutoscaling", +"description": "The desired config for pod autoscaling." +}, "desiredPodSecurityPolicyConfig": { "$ref": "PodSecurityPolicyConfig", "description": "The desired configuration options for the PodSecurityPolicy feature." @@ -4655,6 +4667,27 @@ }, "type": "object" }, +"DesiredEnterpriseConfig": { +"description": "DesiredEnterpriseConfig is a wrapper used for updating enterprise_config.", +"id": "DesiredEnterpriseConfig", +"properties": { +"desiredTier": { +"description": "desired_tier specifies the desired tier of the cluster.", +"enum": [ +"CLUSTER_TIER_UNSPECIFIED", +"STANDARD", +"ENTERPRISE" +], +"enumDescriptions": [ +"CLUSTER_TIER_UNSPECIFIED is when cluster_tier is not set.", +"STANDARD indicates a standard GKE cluster.", +"ENTERPRISE indicates a GKE Enterprise cluster." +], +"type": "string" +} +}, +"type": "object" +}, "DnsCacheConfig": { "description": "Configuration for NodeLocal DNSCache", "id": "DnsCacheConfig", @@ -4690,6 +4723,20 @@ ], "readOnly": true, "type": "string" +}, +"desiredTier": { +"description": "desired_tier specifies the desired tier of the cluster.", +"enum": [ +"CLUSTER_TIER_UNSPECIFIED", +"STANDARD", +"ENTERPRISE" +], +"enumDescriptions": [ +"CLUSTER_TIER_UNSPECIFIED is when cluster_tier is not set.", +"STANDARD indicates a standard GKE cluster.", +"ENTERPRISE indicates a GKE Enterprise cluster." +], +"type": "string" } }, "type": "object" @@ -7109,6 +7156,27 @@ false }, "type": "object" }, +"PodAutoscaling": { +"description": "PodAutoscaling is used for configuration of parameters for workload autoscaling.", +"id": "PodAutoscaling", +"properties": { +"hpaProfile": { +"description": "Selected Horizontal Pod Autoscaling profile.", +"enum": [ +"HPA_PROFILE_UNSPECIFIED", +"NONE", +"PERFORMANCE" +], +"enumDescriptions": [ +"HPA_PROFILE_UNSPECIFIED is used when no custom HPA profile is set.", +"Customers explicitly opt-out of HPA profiles.", +"PERFORMANCE is used when customers opt-in to the performance HPA profile. In this profile we support a higher number of HPAs per cluster and faster metrics collection for workload autoscaling." +], +"type": "string" +} +}, +"type": "object" +}, "PodCIDROverprovisionConfig": { "description": "[PRIVATE FIELD] Config for pod CIDR size overprovisioning.", "id": "PodCIDROverprovisionConfig", diff --git a/googleapiclient/discovery_cache/documents/containeranalysis.v1alpha1.json b/googleapiclient/discovery_cache/documents/containeranalysis.v1alpha1.json index cac2aee9b73..dc095b694cb 100644 --- a/googleapiclient/discovery_cache/documents/containeranalysis.v1alpha1.json +++ b/googleapiclient/discovery_cache/documents/containeranalysis.v1alpha1.json @@ -1420,7 +1420,7 @@ } } }, -"revision": "20240924", +"revision": "20241025", "rootUrl": "https://containeranalysis.googleapis.com/", "schemas": { "AnalysisCompleted": { @@ -4616,11 +4616,13 @@ false "description": "The type of hash that was performed.", "enum": [ "NONE", -"SHA256" +"SHA256", +"GO_MODULE_H1" ], "enumDescriptions": [ "No hash requested.", -"A sha256 hash." +"A sha256 hash.", +"Dirhash of a Go module's source code which is then hex-encoded." ], "type": "string" }, diff --git a/googleapiclient/discovery_cache/documents/dataflow.v1b3.json b/googleapiclient/discovery_cache/documents/dataflow.v1b3.json index bc5177e4d4e..ed5f6b2bf0e 100644 --- a/googleapiclient/discovery_cache/documents/dataflow.v1b3.json +++ b/googleapiclient/discovery_cache/documents/dataflow.v1b3.json @@ -2208,7 +2208,7 @@ } } }, -"revision": "20241014", +"revision": "20241028", "rootUrl": "https://dataflow.googleapis.com/", "schemas": { "ApproximateProgress": { @@ -3632,6 +3632,34 @@ }, "type": "object" }, +"GPUUsage": { +"description": "Information about the GPU usage on the worker.", +"id": "GPUUsage", +"properties": { +"timestamp": { +"description": "Required. Timestamp of the measurement.", +"format": "google-datetime", +"type": "string" +}, +"utilization": { +"$ref": "GPUUtilization", +"description": "Required. Utilization info about the GPU." +} +}, +"type": "object" +}, +"GPUUtilization": { +"description": "Utilization details about the GPU.", +"id": "GPUUtilization", +"properties": { +"rate": { +"description": "Required. GPU utilization rate of any kernel over the last sample period in the range of [0, 1].", +"format": "double", +"type": "number" +} +}, +"type": "object" +}, "GetDebugConfigRequest": { "description": "Request to get updated debug configuration for component.", "id": "GetDebugConfigRequest", @@ -5468,6 +5496,13 @@ false }, "type": "array" }, +"gpuUsage": { +"description": "Optional. GPU usage samples.", +"items": { +"$ref": "GPUUsage" +}, +"type": "array" +}, "memoryInfo": { "description": "Memory utilization samples.", "items": { diff --git a/googleapiclient/discovery_cache/documents/datastream.v1.json b/googleapiclient/discovery_cache/documents/datastream.v1.json index 9a38717a7ac..b1d58823c9c 100644 --- a/googleapiclient/discovery_cache/documents/datastream.v1.json +++ b/googleapiclient/discovery_cache/documents/datastream.v1.json @@ -1250,7 +1250,7 @@ } } }, -"revision": "20240930", +"revision": "20241028", "rootUrl": "https://datastream.googleapis.com/", "schemas": { "AppendOnly": { @@ -1394,6 +1394,27 @@ "properties": {}, "type": "object" }, +"BinaryLogParser": { +"description": "Configuration to use Binary Log Parser CDC technique.", +"id": "BinaryLogParser", +"properties": { +"logFileDirectories": { +"$ref": "LogFileDirectories", +"description": "Use Oracle directories." +}, +"oracleAsmLogFileAccess": { +"$ref": "OracleAsmLogFileAccess", +"description": "Use Oracle ASM." +} +}, +"type": "object" +}, +"BinaryLogPosition": { +"description": "Use Binary log position based replication.", +"id": "BinaryLogPosition", +"properties": {}, +"type": "object" +}, "CancelOperationRequest": { "description": "The request message for Operations.CancelOperation.", "id": "CancelOperationRequest", @@ -1724,6 +1745,12 @@ }, "type": "object" }, +"Gtid": { +"description": "Use GTID based replication.", +"id": "Gtid", +"properties": {}, +"type": "object" +}, "JsonFileFormat": { "description": "JSON file format configuration.", "id": "JsonFileFormat", @@ -1947,6 +1974,27 @@ }, "type": "object" }, +"LogFileDirectories": { +"description": "Configuration to specify the Oracle directories to access the log files.", +"id": "LogFileDirectories", +"properties": { +"archivedLogDirectory": { +"description": "Required. Oracle directory for archived logs.", +"type": "string" +}, +"onlineLogDirectory": { +"description": "Required. Oracle directory for online logs.", +"type": "string" +} +}, +"type": "object" +}, +"LogMiner": { +"description": "Configuration to use LogMiner CDC method.", +"id": "LogMiner", +"properties": {}, +"type": "object" +}, "LookupStreamObjectRequest": { "description": "Request for looking up a specific stream object by its source object identifier.", "id": "LookupStreamObjectRequest", @@ -2112,10 +2160,18 @@ "description": "MySQL source configuration", "id": "MysqlSourceConfig", "properties": { +"binaryLogPosition": { +"$ref": "BinaryLogPosition", +"description": "Use Binary log position based replication." +}, "excludeObjects": { "$ref": "MysqlRdbms", "description": "MySQL objects to exclude from the stream." }, +"gtid": { +"$ref": "Gtid", +"description": "Use GTID based replication." +}, "includeObjects": { "$ref": "MysqlRdbms", "description": "MySQL objects to retrieve from the source." @@ -2275,6 +2331,51 @@ }, "type": "object" }, +"OracleAsmConfig": { +"description": "Configuration for Oracle Automatic Storage Management (ASM) connection.", +"id": "OracleAsmConfig", +"properties": { +"asmService": { +"description": "Required. ASM service name for the Oracle ASM connection.", +"type": "string" +}, +"connectionAttributes": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. Connection string attributes", +"type": "object" +}, +"hostname": { +"description": "Required. Hostname for the Oracle ASM connection.", +"type": "string" +}, +"oracleSslConfig": { +"$ref": "OracleSslConfig", +"description": "Optional. SSL configuration for the Oracle connection." +}, +"password": { +"description": "Required. Password for the Oracle ASM connection.", +"type": "string" +}, +"port": { +"description": "Required. Port for the Oracle ASM connection.", +"format": "int32", +"type": "integer" +}, +"username": { +"description": "Required. Username for the Oracle ASM connection.", +"type": "string" +} +}, +"type": "object" +}, +"OracleAsmLogFileAccess": { +"description": "Configuration to use Oracle ASM to access the log files.", +"id": "OracleAsmLogFileAccess", +"properties": {}, +"type": "object" +}, "OracleColumn": { "description": "Oracle Column.", "id": "OracleColumn", @@ -2356,6 +2457,10 @@ "description": "Required. Hostname for the Oracle connection.", "type": "string" }, +"oracleAsmConfig": { +"$ref": "OracleAsmConfig", +"description": "Optional. Configuration for Oracle ASM connection." +}, "oracleSslConfig": { "$ref": "OracleSslConfig", "description": "Optional. SSL configuration for the Oracle connection." @@ -2369,6 +2474,10 @@ "format": "int32", "type": "integer" }, +"secretManagerStoredPassword": { +"description": "Optional. A reference to a Secret Manager resource name storing the Oracle connection password. Mutually exclusive with the `password` field.", +"type": "string" +}, "username": { "description": "Required. Username for the Oracle connection.", "type": "string" @@ -2424,6 +2533,10 @@ "description": "Oracle data source configuration", "id": "OracleSourceConfig", "properties": { +"binaryLogParser": { +"$ref": "BinaryLogParser", +"description": "Use Binary Log Parser." +}, "dropLargeObjects": { "$ref": "DropLargeObjects", "description": "Drop large object values." @@ -2436,6 +2549,10 @@ "$ref": "OracleRdbms", "description": "Oracle objects to include in the stream." }, +"logMiner": { +"$ref": "LogMiner", +"description": "Use LogMiner." +}, "maxConcurrentBackfillTasks": { "description": "Maximum number of concurrent backfill tasks. The number should be non-negative. If not set (or set to 0), the system's default value is used.", "format": "int32", @@ -2546,7 +2663,7 @@ "type": "object" }, "PostgresqlProfile": { -"description": "PostgreSQL database profile. Next ID: 7.", +"description": "PostgreSQL database profile.", "id": "PostgresqlProfile", "properties": { "database": { diff --git a/googleapiclient/discovery_cache/documents/developerconnect.v1.json b/googleapiclient/discovery_cache/documents/developerconnect.v1.json index 9fa9727373a..c5c7253b29d 100644 --- a/googleapiclient/discovery_cache/documents/developerconnect.v1.json +++ b/googleapiclient/discovery_cache/documents/developerconnect.v1.json @@ -808,10 +808,7 @@ }, "response": { "$ref": "Empty" -}, -"scopes": [ -"https://www.googleapis.com/auth/cloud-platform" -] +} }, "processGitLabWebhook": { "description": "ProcessGitLabWebhook is called by the GitLab.com for notifying events.", @@ -970,7 +967,7 @@ } } }, -"revision": "20241017", +"revision": "20241023", "rootUrl": "https://developerconnect.googleapis.com/", "schemas": { "CancelOperationRequest": { diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json index 32c3e209aaa..459a770ba3a 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v2beta1.json @@ -7822,6 +7822,163 @@ } } }, +"sipTrunks": { +"methods": { +"create": { +"description": "Creates a SipTrunk for a specified location.", +"flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/sipTrunks", +"httpMethod": "POST", +"id": "dialogflow.projects.locations.sipTrunks.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. The location to create a SIP trunk for. Format: `projects//locations/`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2beta1/{+parent}/sipTrunks", +"request": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"response": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/dialogflow" +] +}, +"delete": { +"description": "Deletes a specified SipTrunk.", +"flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/sipTrunks/{sipTrunksId}", +"httpMethod": "DELETE", +"id": "dialogflow.projects.locations.sipTrunks.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the SIP trunk to delete. Format: `projects//locations//sipTrunks/`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/sipTrunks/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2beta1/{+name}", +"response": { +"$ref": "GoogleProtobufEmpty" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/dialogflow" +] +}, +"get": { +"description": "Retrieves the specified SipTrunk.", +"flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/sipTrunks/{sipTrunksId}", +"httpMethod": "GET", +"id": "dialogflow.projects.locations.sipTrunks.get", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the SIP trunk to delete. Format: `projects//locations//sipTrunks/`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/sipTrunks/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2beta1/{+name}", +"response": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/dialogflow" +] +}, +"list": { +"description": "Returns a list of SipTrunks in the specified location.", +"flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/sipTrunks", +"httpMethod": "GET", +"id": "dialogflow.projects.locations.sipTrunks.list", +"parameterOrder": [ +"parent" +], +"parameters": { +"pageSize": { +"description": "Optional. The maximum number of items to return in a single page. By default 100 and at most 1000.", +"format": "int32", +"location": "query", +"type": "integer" +}, +"pageToken": { +"description": "Optional. The next_page_token value returned from a previous list request.", +"location": "query", +"type": "string" +}, +"parent": { +"description": "Required. The location to list SIP trunks from. Format: `projects//locations/`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2beta1/{+parent}/sipTrunks", +"response": { +"$ref": "GoogleCloudDialogflowV2beta1ListSipTrunksResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/dialogflow" +] +}, +"patch": { +"description": "Updates the specified SipTrunk.", +"flatPath": "v2beta1/projects/{projectsId}/locations/{locationsId}/sipTrunks/{sipTrunksId}", +"httpMethod": "PATCH", +"id": "dialogflow.projects.locations.sipTrunks.patch", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/sipTrunks/[^/]+$", +"required": true, +"type": "string" +}, +"updateMask": { +"description": "Optional. The mask to control which fields get updated. If the mask is not present, all fields will be updated.", +"format": "google-fieldmask", +"location": "query", +"type": "string" +} +}, +"path": "v2beta1/{+name}", +"request": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"response": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/dialogflow" +] +} +} +}, "statelessSuggestion": { "methods": { "generate": { @@ -8082,7 +8239,7 @@ } } }, -"revision": "20241019", +"revision": "20241025", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -16021,6 +16178,86 @@ true "properties": {}, "type": "object" }, +"GoogleCloudDialogflowV2beta1Connection": { +"description": "Represents a connection for SIP Trunk.", +"id": "GoogleCloudDialogflowV2beta1Connection", +"properties": { +"connectionId": { +"description": "Output only. The unique identifier of the SIP Trunk connection.", +"readOnly": true, +"type": "string" +}, +"errorDetails": { +"$ref": "GoogleCloudDialogflowV2beta1ConnectionErrorDetails", +"description": "Output only. The error details for the connection. Only populated when authentication errors occur.", +"readOnly": true +}, +"state": { +"description": "Output only. State of the connection.", +"enum": [ +"STATE_UNSPECIFIED", +"CONNECTED", +"DISCONNECTED", +"AUTHENTICATION_FAILED", +"KEEPALIVE" +], +"enumDescriptions": [ +"SIP Trunk connection state is Not specified.", +"SIP Trunk connection is connected.", +"SIP Trunk connection is disconnected.", +"SIP Trunk connection has authentication error.", +"SIP Trunk connection is keepalive." +], +"readOnly": true, +"type": "string" +}, +"updateTime": { +"description": "Output only. When the connection status changed.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleCloudDialogflowV2beta1ConnectionErrorDetails": { +"description": "The error details of Sip Trunk connection authentication.", +"id": "GoogleCloudDialogflowV2beta1ConnectionErrorDetails", +"properties": { +"certificateState": { +"description": "Output only. The status of the certificate authentication.", +"enum": [ +"CERTIFICATE_STATE_UNSPECIFIED", +"CERTIFICATE_VALID", +"CERTIFICATE_INVALID", +"CERTIFICATE_EXPIRED", +"CERTIFICATE_HOSTNAME_NOT_FOUND", +"CERTIFICATE_UNAUTHENTICATED", +"CERTIFICATE_TRUST_STORE_NOT_FOUND", +"CERTIFICATE_HOSTNAME_INVALID_FORMAT", +"CERTIFICATE_QUOTA_EXCEEDED" +], +"enumDescriptions": [ +"Certificate state is not specified.", +"Certificate is valid.", +"Catch all for any error not specified.", +"Certificate leaf node has expired.", +"There is no hostname defined to authenticate in SipTrunkingServer.", +"No path found from the leaf certificate to any root.", +"Trust store does not exist.", +"Hostname has invalid format.", +"Certificate has exhausted its quota." +], +"readOnly": true, +"type": "string" +}, +"errorMessage": { +"description": "The error message provided from SIP trunking auth service", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDialogflowV2beta1Context": { "description": "Dialogflow contexts are similar to natural language context. If a person says to you \"they are orange\", you need context in order to understand what \"they\" is referring to. Similarly, for Dialogflow to handle an end-user expression like that, it needs to be provided with context in order to correctly match an intent. Using contexts, you can control the flow of a conversation. You can configure contexts for an intent by setting input and output contexts, which are identified by string names. When an intent is matched, any configured output contexts for that intent become active. While any contexts are active, Dialogflow is more likely to match intents that are configured with input contexts that correspond to the currently active contexts. For more information about context, see the [Contexts guide](https://cloud.google.com/dialogflow/docs/contexts-overview).", "id": "GoogleCloudDialogflowV2beta1Context", @@ -19411,6 +19648,24 @@ true }, "type": "object" }, +"GoogleCloudDialogflowV2beta1ListSipTrunksResponse": { +"description": "The response message for SipTrunks.ListSipTrunks.", +"id": "GoogleCloudDialogflowV2beta1ListSipTrunksResponse", +"properties": { +"nextPageToken": { +"description": "Token to retrieve the next page of results, or empty if there are no more results in the list.", +"type": "string" +}, +"sipTrunks": { +"description": "The list of SIP trunks.", +"items": { +"$ref": "GoogleCloudDialogflowV2beta1SipTrunk" +}, +"type": "array" +} +}, +"type": "object" +}, "GoogleCloudDialogflowV2beta1ListSuggestionsResponse": { "deprecated": true, "description": "The response message for Participants.ListSuggestions.", @@ -20530,6 +20785,36 @@ true }, "type": "object" }, +"GoogleCloudDialogflowV2beta1SipTrunk": { +"description": "SipTrunk is the resource that represents a SIP trunk to connect to Google Telephony platform SIP trunking service.", +"id": "GoogleCloudDialogflowV2beta1SipTrunk", +"properties": { +"connections": { +"description": "Output only. Connections of the SIP trunk.", +"items": { +"$ref": "GoogleCloudDialogflowV2beta1Connection" +}, +"readOnly": true, +"type": "array" +}, +"displayName": { +"description": "Optional. Human readable alias for this trunk.", +"type": "string" +}, +"expectedHostname": { +"description": "Required. The expected hostnames in the peer certificate from partner that is used for TLS authentication.", +"items": { +"type": "string" +}, +"type": "array" +}, +"name": { +"description": "Identifier. The unique identifier of the SIP trunk. Format: `projects//locations//sipTrunks/`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDialogflowV2beta1SmartReplyAnswer": { "description": "Represents a smart reply answer.", "id": "GoogleCloudDialogflowV2beta1SmartReplyAnswer", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3.json b/googleapiclient/discovery_cache/documents/dialogflow.v3.json index 3454147bb27..d8eb9493ff5 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3.json @@ -4453,7 +4453,7 @@ } } }, -"revision": "20241019", +"revision": "20241025", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -6914,6 +6914,10 @@ "description": "Required. The human-readable name of the generator, unique within the agent. The prompt contains pre-defined parameters such as $conversation, $last-user-utterance, etc. populated by Dialogflow. It can also contain custom placeholders which will be resolved during fulfillment.", "type": "string" }, +"modelParameter": { +"$ref": "GoogleCloudDialogflowCxV3GeneratorModelParameter", +"description": "Parameters passed to the LLM to configure its behavior." +}, "name": { "description": "The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`.", "type": "string" @@ -6932,6 +6936,33 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3GeneratorModelParameter": { +"description": "Parameters to be passed to the LLM. If not set, default values will be used.", +"id": "GoogleCloudDialogflowCxV3GeneratorModelParameter", +"properties": { +"maxDecodeSteps": { +"description": "The maximum number of tokens to generate.", +"format": "int32", +"type": "integer" +}, +"temperature": { +"description": "The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random.", +"format": "float", +"type": "number" +}, +"topK": { +"description": "If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random.", +"format": "int32", +"type": "integer" +}, +"topP": { +"description": "If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random.", +"format": "float", +"type": "number" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3GeneratorPlaceholder": { "description": "Represents a custom placeholder in the prompt text.", "id": "GoogleCloudDialogflowCxV3GeneratorPlaceholder", diff --git a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json index 10fb984ba2d..f381ddfbcbe 100644 --- a/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json +++ b/googleapiclient/discovery_cache/documents/dialogflow.v3beta1.json @@ -5185,7 +5185,7 @@ } } }, -"revision": "20241019", +"revision": "20241025", "rootUrl": "https://dialogflow.googleapis.com/", "schemas": { "GoogleCloudDialogflowCxV3AdvancedSettings": { @@ -10465,6 +10465,10 @@ "$ref": "GoogleCloudDialogflowCxV3beta1LlmModelSettings", "description": "The LLM model settings." }, +"modelParameter": { +"$ref": "GoogleCloudDialogflowCxV3beta1GeneratorModelParameter", +"description": "Parameters passed to the LLM to configure its behavior." +}, "name": { "description": "The unique identifier of the generator. Must be set for the Generators.UpdateGenerator method. Generators.CreateGenerate populates the name automatically. Format: `projects//locations//agents//generators/`.", "type": "string" @@ -10483,6 +10487,33 @@ }, "type": "object" }, +"GoogleCloudDialogflowCxV3beta1GeneratorModelParameter": { +"description": "Parameters to be passed to the LLM. If not set, default values will be used.", +"id": "GoogleCloudDialogflowCxV3beta1GeneratorModelParameter", +"properties": { +"maxDecodeSteps": { +"description": "The maximum number of tokens to generate.", +"format": "int32", +"type": "integer" +}, +"temperature": { +"description": "The temperature used for sampling. Temperature sampling occurs after both topP and topK have been applied. Valid range: [0.0, 1.0] Low temperature = less random. High temperature = more random.", +"format": "float", +"type": "number" +}, +"topK": { +"description": "If set, the sampling process in each step is limited to the top_k tokens with highest probabilities. Valid range: [1, 40] or 1000+. Small topK = less random. Large topK = more random.", +"format": "int32", +"type": "integer" +}, +"topP": { +"description": "If set, only the tokens comprising the top top_p probability mass are considered. If both top_p and top_k are set, top_p will be used for further refining candidates selected with top_k. Valid range: (0.0, 1.0]. Small topP = less random. Large topP = more random.", +"format": "float", +"type": "number" +} +}, +"type": "object" +}, "GoogleCloudDialogflowCxV3beta1GeneratorPlaceholder": { "description": "Represents a custom placeholder in the prompt text.", "id": "GoogleCloudDialogflowCxV3beta1GeneratorPlaceholder", diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1.json index 69663127f27..200df1fe2bb 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1.json @@ -6090,7 +6090,7 @@ } } }, -"revision": "20241025", +"revision": "20241030", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "GoogleApiDistribution": { @@ -13032,6 +13032,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata", @@ -13373,6 +13390,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata", @@ -13961,6 +13995,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1alphaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaFieldConfig": { "description": "Configurations for fields of a schema. For example, configuring a field is indexable, or searchable.", "id": "GoogleCloudDiscoveryengineV1alphaFieldConfig", @@ -15842,6 +15901,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1alphaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaTargetSite": { "description": "A target site for the SiteSearchEngine.", "id": "GoogleCloudDiscoveryengineV1alphaTargetSite", @@ -16466,6 +16547,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata", @@ -16772,6 +16870,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata", @@ -17213,6 +17328,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1betaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaImportCompletionSuggestionsMetadata": { "description": "Metadata related to the progress of the ImportCompletionSuggestions operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaImportCompletionSuggestionsMetadata", @@ -18524,6 +18664,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1betaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaTargetSite": { "description": "A target site for the SiteSearchEngine.", "id": "GoogleCloudDiscoveryengineV1betaTargetSite", diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json index 80d1b19ebbf..1ce3fe5fb60 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1alpha.json @@ -2840,6 +2840,63 @@ } } }, +"sitemaps": { +"methods": { +"create": { +"description": "Creates a Sitemap.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/collections/{collectionsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps", +"httpMethod": "POST", +"id": "discoveryengine.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/collections/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+parent}/sitemaps", +"request": { +"$ref": "GoogleCloudDiscoveryengineV1alphaSitemap" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"delete": { +"description": "Deletes a Sitemap.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/collections/{collectionsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps/{sitemapsId}", +"httpMethod": "DELETE", +"id": "discoveryengine.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/collections/[^/]+/dataStores/[^/]+/siteSearchEngine/sitemaps/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +}, "targetSites": { "methods": { "batchCreate": { @@ -6490,6 +6547,94 @@ } }, "resources": { +"sitemaps": { +"methods": { +"create": { +"description": "Creates a Sitemap.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps", +"httpMethod": "POST", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+parent}/sitemaps", +"request": { +"$ref": "GoogleCloudDiscoveryengineV1alphaSitemap" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"delete": { +"description": "Deletes a Sitemap.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps/{sitemapsId}", +"httpMethod": "DELETE", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine/sitemaps/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"fetch": { +"description": "Fetch Sitemaps in a DataStore.", +"flatPath": "v1alpha/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps:fetch", +"httpMethod": "GET", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.fetch", +"parameterOrder": [ +"parent" +], +"parameters": { +"matcher.urisMatcher.uris": { +"description": "The Sitemap uris.", +"location": "query", +"repeated": true, +"type": "string" +}, +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1alpha/{+parent}/sitemaps:fetch", +"response": { +"$ref": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +}, "targetSites": { "methods": { "batchCreate": { @@ -7862,7 +8007,7 @@ } } }, -"revision": "20241025", +"revision": "20241030", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "GoogleApiDistribution": { @@ -12303,6 +12448,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata", @@ -12681,6 +12843,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata", @@ -13563,6 +13742,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1alphaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaFhirStoreSource": { "description": "Cloud FhirStore source import data from.", "id": "GoogleCloudDiscoveryengineV1alphaFhirStoreSource", @@ -17591,6 +17795,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1alphaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaSpannerSource": { "description": "The Spanner source for importing data", "id": "GoogleCloudDiscoveryengineV1alphaSpannerSource", @@ -18469,6 +18695,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata", @@ -18775,6 +19018,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata", @@ -19216,6 +19476,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1betaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaImportCompletionSuggestionsMetadata": { "description": "Metadata related to the progress of the ImportCompletionSuggestions operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaImportCompletionSuggestionsMetadata", @@ -20527,6 +20812,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1betaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaTargetSite": { "description": "A target site for the SiteSearchEngine.", "id": "GoogleCloudDiscoveryengineV1betaTargetSite", diff --git a/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json b/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json index 94501db3ff6..2e9e96d9fe8 100644 --- a/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json +++ b/googleapiclient/discovery_cache/documents/discoveryengine.v1beta.json @@ -2325,6 +2325,63 @@ } } }, +"sitemaps": { +"methods": { +"create": { +"description": "Creates a Sitemap.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/collections/{collectionsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps", +"httpMethod": "POST", +"id": "discoveryengine.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/collections/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+parent}/sitemaps", +"request": { +"$ref": "GoogleCloudDiscoveryengineV1betaSitemap" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"delete": { +"description": "Deletes a Sitemap.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/collections/{collectionsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps/{sitemapsId}", +"httpMethod": "DELETE", +"id": "discoveryengine.projects.locations.collections.dataStores.siteSearchEngine.sitemaps.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/collections/[^/]+/dataStores/[^/]+/siteSearchEngine/sitemaps/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +}, "targetSites": { "methods": { "batchCreate": { @@ -5794,6 +5851,94 @@ } }, "resources": { +"sitemaps": { +"methods": { +"create": { +"description": "Creates a Sitemap.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps", +"httpMethod": "POST", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+parent}/sitemaps", +"request": { +"$ref": "GoogleCloudDiscoveryengineV1betaSitemap" +}, +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"delete": { +"description": "Deletes a Sitemap.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps/{sitemapsId}", +"httpMethod": "DELETE", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.delete", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. Full resource name of Sitemap, such as `projects/{project}/locations/{location}/collections/{collection}/dataStores/{data_store}/siteSearchEngine/sitemaps/{sitemap}`. If the caller does not have permission to access the Sitemap, regardless of whether or not it exists, a PERMISSION_DENIED error is returned. If the requested Sitemap does not exist, a NOT_FOUND error is returned.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine/sitemaps/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+name}", +"response": { +"$ref": "GoogleLongrunningOperation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, +"fetch": { +"description": "Fetch Sitemaps in a DataStore.", +"flatPath": "v1beta/projects/{projectsId}/locations/{locationsId}/dataStores/{dataStoresId}/siteSearchEngine/sitemaps:fetch", +"httpMethod": "GET", +"id": "discoveryengine.projects.locations.dataStores.siteSearchEngine.sitemaps.fetch", +"parameterOrder": [ +"parent" +], +"parameters": { +"matcher.urisMatcher.uris": { +"description": "The Sitemap uris.", +"location": "query", +"repeated": true, +"type": "string" +}, +"parent": { +"description": "Required. Parent resource name of the SiteSearchEngine, such as `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine`.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/dataStores/[^/]+/siteSearchEngine$", +"required": true, +"type": "string" +} +}, +"path": "v1beta/{+parent}/sitemaps:fetch", +"response": { +"$ref": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +} +} +}, "targetSites": { "methods": { "batchCreate": { @@ -7060,7 +7205,7 @@ } } }, -"revision": "20241025", +"revision": "20241030", "rootUrl": "https://discoveryengine.googleapis.com/", "schemas": { "GoogleApiDistribution": { @@ -9921,6 +10066,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaCreateTargetSiteMetadata", @@ -10262,6 +10424,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1alphaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1alphaDeleteTargetSiteMetadata", @@ -10850,6 +11029,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1alphaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1alphaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaFieldConfig": { "description": "Configurations for fields of a schema. For example, configuring a field is indexable, or searchable.", "id": "GoogleCloudDiscoveryengineV1alphaFieldConfig", @@ -12731,6 +12935,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1alphaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1alphaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1alphaTargetSite": { "description": "A target site for the SiteSearchEngine.", "id": "GoogleCloudDiscoveryengineV1alphaTargetSite", @@ -15326,6 +15552,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.CreateSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaCreateSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.CreateTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaCreateTargetSiteMetadata", @@ -15669,6 +15912,23 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata": { +"description": "Metadata related to the progress of the SiteSearchEngineService.DeleteSitemap operation. This will be returned by the google.longrunning.Operation.metadata field.", +"id": "GoogleCloudDiscoveryengineV1betaDeleteSitemapMetadata", +"properties": { +"createTime": { +"description": "Operation create time.", +"format": "google-datetime", +"type": "string" +}, +"updateTime": { +"description": "Operation last update time. If the operation is done, this is also the finish time.", +"format": "google-datetime", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata": { "description": "Metadata related to the progress of the SiteSearchEngineService.DeleteTargetSite operation. This will be returned by the google.longrunning.Operation.metadata field.", "id": "GoogleCloudDiscoveryengineV1betaDeleteTargetSiteMetadata", @@ -16305,6 +16565,31 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse": { +"description": "Response message for SiteSearchEngineService.FetchSitemaps method.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponse", +"properties": { +"sitemapsMetadata": { +"description": "List of Sitemaps fetched.", +"items": { +"$ref": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata": { +"description": "Contains a Sitemap and its metadata.", +"id": "GoogleCloudDiscoveryengineV1betaFetchSitemapsResponseSitemapMetadata", +"properties": { +"sitemap": { +"$ref": "GoogleCloudDiscoveryengineV1betaSitemap", +"description": "The Sitemap." +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaFhirStoreSource": { "description": "Cloud FhirStore source import data from.", "id": "GoogleCloudDiscoveryengineV1betaFhirStoreSource", @@ -19577,6 +19862,28 @@ }, "type": "object" }, +"GoogleCloudDiscoveryengineV1betaSitemap": { +"description": "A sitemap for the SiteSearchEngine.", +"id": "GoogleCloudDiscoveryengineV1betaSitemap", +"properties": { +"createTime": { +"description": "Output only. The sitemap's creation time.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The fully qualified resource name of the sitemap. `projects/*/locations/*/collections/*/dataStores/*/siteSearchEngine/sitemaps/*` The `sitemap_id` suffix is system-generated.", +"readOnly": true, +"type": "string" +}, +"uri": { +"description": "Public URI for the sitemap, e.g. `www.example.com/sitemap.xml`.", +"type": "string" +} +}, +"type": "object" +}, "GoogleCloudDiscoveryengineV1betaSpannerSource": { "description": "The Spanner source for importing data", "id": "GoogleCloudDiscoveryengineV1betaSpannerSource", diff --git a/googleapiclient/discovery_cache/documents/displayvideo.v2.json b/googleapiclient/discovery_cache/documents/displayvideo.v2.json index 9d8d28e3cec..2bb286b9253 100644 --- a/googleapiclient/discovery_cache/documents/displayvideo.v2.json +++ b/googleapiclient/discovery_cache/documents/displayvideo.v2.json @@ -9267,7 +9267,7 @@ } } }, -"revision": "20241024", +"revision": "20241031", "rootUrl": "https://displayvideo.googleapis.com/", "schemas": { "ActivateManualTriggerRequest": { @@ -14113,7 +14113,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -14198,7 +14199,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" } @@ -14308,7 +14310,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -14393,7 +14396,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -14504,7 +14508,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -14589,7 +14594,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -14701,7 +14707,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -14786,7 +14793,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "readOnly": true, "type": "string" @@ -15749,7 +15757,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -15834,7 +15843,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -16662,7 +16672,8 @@ true "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -16747,7 +16758,8 @@ true "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, diff --git a/googleapiclient/discovery_cache/documents/displayvideo.v3.json b/googleapiclient/discovery_cache/documents/displayvideo.v3.json index 8643a74aa13..8280fe8422c 100644 --- a/googleapiclient/discovery_cache/documents/displayvideo.v3.json +++ b/googleapiclient/discovery_cache/documents/displayvideo.v3.json @@ -9222,7 +9222,7 @@ } } }, -"revision": "20241024", +"revision": "20241031", "rootUrl": "https://displayvideo.googleapis.com/", "schemas": { "ActiveViewVideoViewabilityMetricConfig": { @@ -10320,7 +10320,8 @@ "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -10405,7 +10406,8 @@ "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -15100,7 +15102,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -15185,7 +15188,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" } @@ -15295,7 +15299,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -15380,7 +15385,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -15491,7 +15497,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -15576,7 +15583,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -15688,7 +15696,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -15773,7 +15782,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "readOnly": true, "type": "string" @@ -16736,7 +16746,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -16821,7 +16832,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, @@ -17663,7 +17675,8 @@ false "EXCHANGE_CHARTBOOST_GBID", "EXCHANGE_ADMOST_GBID", "EXCHANGE_TOPON_GBID", -"EXCHANGE_NETFLIX" +"EXCHANGE_NETFLIX", +"EXCHANGE_TUBI" ], "enumDescriptions": [ "Exchange is not specified or is unknown in this version.", @@ -17748,7 +17761,8 @@ false "Chartboost Mediation.", "AdMost.", "TopOn.", -"Netflix." +"Netflix.", +"Tubi." ], "type": "string" }, diff --git a/googleapiclient/discovery_cache/documents/dlp.v2.json b/googleapiclient/discovery_cache/documents/dlp.v2.json index 6bea2d9f22b..dc804ed2128 100644 --- a/googleapiclient/discovery_cache/documents/dlp.v2.json +++ b/googleapiclient/discovery_cache/documents/dlp.v2.json @@ -325,7 +325,7 @@ "infoTypes": { "methods": { "list": { -"description": "Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.", +"description": "Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.", "flatPath": "v2/infoTypes", "httpMethod": "GET", "id": "dlp.infoTypes.list", @@ -367,7 +367,7 @@ "infoTypes": { "methods": { "list": { -"description": "Returns a list of the sensitive information types that DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.", +"description": "Returns a list of the sensitive information types that the DLP API supports. See https://cloud.google.com/sensitive-data-protection/docs/infotypes-reference to learn more.", "flatPath": "v2/locations/{locationsId}/infoTypes", "httpMethod": "GET", "id": "dlp.locations.infoTypes.list", @@ -5015,7 +5015,7 @@ } } }, -"revision": "20241020", +"revision": "20241027", "rootUrl": "https://dlp.googleapis.com/", "schemas": { "GooglePrivacyDlpV2Action": { @@ -5771,7 +5771,7 @@ "type": "string" }, "maxConnections": { -"description": "Required. DLP will limit its connections to max_connections. Must be 2 or greater.", +"description": "Required. The DLP API will limit its connections to max_connections. Must be 2 or greater.", "format": "int32", "type": "integer" }, @@ -6209,7 +6209,7 @@ "type": "object" }, "GooglePrivacyDlpV2Connection": { -"description": "A data connection to allow DLP to profile data in locations that require additional configuration.", +"description": "A data connection to allow the DLP API to profile data in locations that require additional configuration.", "id": "GooglePrivacyDlpV2Connection", "properties": { "cloudSql": { @@ -6239,7 +6239,7 @@ ], "enumDescriptions": [ "Unused", -"DLP automatically created this connection during an initial scan, and it is awaiting full configuration by a user.", +"The DLP API automatically created this connection during an initial scan, and it is awaiting full configuration by a user.", "A configured connection that has not encountered any errors.", "A configured connection that encountered errors during its last use. It will not be used again until it is set to AVAILABLE. If the resolution requires external action, then the client must send a request to set the status to AVAILABLE when the connection is ready for use. If the resolution doesn't require external action, then any changes to the connection properties will automatically mark it as AVAILABLE." ], @@ -6647,7 +6647,7 @@ }, "publishToScc": { "$ref": "GooglePrivacyDlpV2PublishToSecurityCommandCenter", -"description": "Publishes findings to SCC for each data profile." +"description": "Publishes findings to Security Command Center for each data profile." }, "tagResources": { "$ref": "GooglePrivacyDlpV2TagResources", @@ -6731,7 +6731,7 @@ "description": "Must be set only when scanning other clouds." }, "projectId": { -"description": "The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled.", +"description": "The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled.", "type": "string" } }, @@ -10226,7 +10226,7 @@ "description": "The data to scan: folder, org, or project" }, "projectId": { -"description": "The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the Cloud DLP API must be enabled.", +"description": "The project that will run the scan. The DLP service account that exists within this project must have access to all resources that are profiled, and the DLP API must be enabled.", "type": "string" } }, @@ -10706,7 +10706,7 @@ "type": "object" }, "GooglePrivacyDlpV2PublishToSecurityCommandCenter": { -"description": "If set, a summary finding will be created/updated in SCC for each profile.", +"description": "If set, a summary finding will be created or updated in Security Command Center for each profile.", "id": "GooglePrivacyDlpV2PublishToSecurityCommandCenter", "properties": {}, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/drive.v3.json b/googleapiclient/discovery_cache/documents/drive.v3.json index 2d0867495b4..8fbd84a00ea 100644 --- a/googleapiclient/discovery_cache/documents/drive.v3.json +++ b/googleapiclient/discovery_cache/documents/drive.v3.json @@ -1764,7 +1764,7 @@ ], "enumDescriptions": [ "Unspecified action", -"The user accepts the proposal", +"The user accepts the proposal. Note: If this action is used, the `role` field must have at least one value.", "The user denies the proposal" ], "location": "query", @@ -2769,7 +2769,7 @@ } } }, -"revision": "20241014", +"revision": "20241027", "rootUrl": "https://www.googleapis.com/", "schemas": { "About": { diff --git a/googleapiclient/discovery_cache/documents/eventarc.v1.json b/googleapiclient/discovery_cache/documents/eventarc.v1.json index 41c2cf9dfd6..6a8f3917147 100644 --- a/googleapiclient/discovery_cache/documents/eventarc.v1.json +++ b/googleapiclient/discovery_cache/documents/eventarc.v1.json @@ -2369,7 +2369,7 @@ } } }, -"revision": "20241021", +"revision": "20241029", "rootUrl": "https://eventarc.googleapis.com/", "schemas": { "AuditConfig": { @@ -2934,7 +2934,7 @@ "type": "string" }, "serviceAccount": { -"description": "Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests.", +"description": "Required. Service account email used to generate the [OAuth token](https://developers.google.com/identity/protocols/OAuth2). The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow Pipeline to create OAuth2 tokens for authenticated requests.", "type": "string" } }, @@ -2949,7 +2949,7 @@ "type": "string" }, "serviceAccount": { -"description": "Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts?hl=en#sa_common for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests.", +"description": "Required. Service account email used to generate the OIDC Token. The principal who calls this API must have iam.serviceAccounts.actAs permission in the service account. See https://cloud.google.com/iam/docs/understanding-service-accounts for more information. Eventarc service agents must have roles/roles/iam.serviceAccountTokenCreator role to allow the Pipeline to create OpenID tokens for authenticated requests.", "type": "string" } }, @@ -2960,11 +2960,11 @@ "id": "GoogleCloudEventarcV1PipelineDestinationHttpEndpoint", "properties": { "messageBindingTemplate": { -"description": "Optional. The CEL expression used to construct a new HTTP request to be sent to the final destination. The result of the CEL expression must be a map of key/value pairs such that: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type\u2019s canonical representation. If the `headers` field doesn\u2019t exist then the resulting HTTP request will be without headers. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field doesn\u2019t exist then the resulting HTTP request will be without a body. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. Headers added to the request by previous filters in the chain can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers added by previous mediation steps and not the headers present on the original incoming request. For example, the following CEL expression can be used to construct a Headers-only HTTP request by adding an additional header to the headers added by previous mediations in the Pipeline: ``` {\"headers\": headers.merge({\"new-header-key\": \"new-header-value\"})} ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - toMap: list(map).toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn\u2019t then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. If a binding is not specified here, by default the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. The Pipeline converts the data field of the message to the format provided in Pipeline.Destination.output_payload_format and maps it to the body field of the result. It also sets the corresponding Content-Type header to the output_payload_format type. If the Pipeline.Destination.output_payload_format is not set, then the Pipeline will treat the data field of the message as opaque binary data and attach it to the request body as bytes. In this case the Content-type header will be set to the value of the datacontenttype attribute set on the incoming CloudEvent message if present and the `application/octet-stream` MIME type otherwise. The Pipeline expects that the content of the message will adhere to the standard CloudEvent format. If it doesn\u2019t then the outgoing message request may fail with a persistent error.", +"description": "Optional. The CEL expression used to modify how the destination-bound HTTP request is constructed. If a binding expression is not specified here, the message is treated as a CloudEvent and is mapped to the HTTP request according to the CloudEvent HTTP Protocol Binding Binary Content Mode. In this representation, all fields except the `data` and `datacontenttype` field on the message are mapped to HTTP request headers with a prefix of `ce-`. To construct the HTTP request payload and the value of the content-type HTTP header, the payload format is defined as follows: 1) Use the output_payload_format_type on the Pipeline.Destination if it is set, else: 2) Use the input_payload_format_type on the Pipeline if it is set, else: 3) Treat the payload as opaque binary data. The `data` field of the message is converted to the payload format or left as-is for case 3) and then attached as the payload of the HTTP request. The `content-type` header on the HTTP request is set to the payload format type or left empty for case 3). However, if a mediation has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header on the HTTP request is set to this `datacontenttype` value. For example, if the `datacontenttype` is \"application/json\" and the payload format type is \"application/json; charset=utf-8\", then the `content-type` header on the HTTP request is set to \"application/json; charset=utf-8\". If a non-empty binding expression is specified then this expression is used to modify the default CloudEvent HTTP Protocol Binding Binary Content representation. The result of the CEL expression must be a map of key/value pairs which is used as follows: - If a map named `headers` exists on the result of the expression, then its key/value pairs are directly mapped to the HTTP request headers. The headers values are constructed from the corresponding value type's canonical representation. If the `headers` field doesn't exist then the resulting HTTP request will be the headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message. Note: If the specified binding expression, has updated the `datacontenttype` field on the message so that it is not the same as the payload format type but it is still a prefix of the payload format type, then the `content-type` header in the `headers` map is set to this `datacontenttype` value. - If a field named `body` exists on the result of the expression then its value is directly mapped to the body of the request. If the value of the `body` field is of type bytes or string then it is used for the HTTP request body as-is, with no conversion. If the body field is of any other type then it is converted to a JSON string. If the body field does not exist then the resulting payload of the HTTP request will be data value of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. - Any other fields in the resulting expression will be ignored. The CEL expression may access the incoming CloudEvent message in its definition, as follows: - The `data` field of the incoming CloudEvent message can be accessed using the `message.data` value. Subfields of `message.data` may also be accessed if an input_payload_format has been specified on the Pipeline. - Each attribute of the incoming CloudEvent message can be accessed using the `message.` value, where is replaced with the name of the attribute. - Existing headers can be accessed in the CEL expression using the `headers` variable. The `headers` variable defines a map of key/value pairs corresponding to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message as described earlier. For example, the following CEL expression can be used to construct an HTTP request by adding an additional header to the HTTP headers of the CloudEvent HTTP Binding Binary Content Mode representation of the final message and by overwriting the body of the request: ``` { \"headers\": headers.merge({\"new-header-key\": \"new-header-value\"}), \"body\": \"new-body\" } ``` Additionally, the following CEL extension functions are provided for use in this CEL expression: - toBase64Url: map.toBase64Url() -> string - Converts a CelValue to a base64url encoded string - toJsonString: map.toJsonString() -> string - Converts a CelValue to a JSON string - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are \".\" separated indices of the map keys. - For example: { \"a\": 1, \"b\": { \"c\": 2, \"d\": 3 } \"e\": [4, 5] } .denormalize() -> { \"a\": 1, \"b.c\": 2, \"b.d\": 3, \"e.0\": 4, \"e.1\": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function. The Pipeline expects that the message it receives adheres to the standard CloudEvent format. If it doesn't then the outgoing message request may fail with a persistent error.", "type": "string" }, "uri": { -"description": "Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `http://10.10.10.8:80/route`, `http://svc.us-central1.p.local:8080/`. Only HTTP and HTTPS protocols are supported. The host can be either a static IP addressable from the VPC specified by the network config, or an internal DNS hostname of the service resolvable via Cloud DNS.", +"description": "Required. The URI of the HTTP enpdoint. The value must be a RFC2396 URI string. Examples: `https://svc.us-central1.p.local:8080/route`. Only the HTTPS protocol is supported.", "type": "string" } }, @@ -2997,7 +2997,7 @@ "id": "GoogleCloudEventarcV1PipelineMediationTransformation", "properties": { "transformationTemplate": { -"description": "Optional. The template to apply to transform messages.", +"description": "Optional. The CEL expression template to apply to transform messages. The following CEL extension functions are provided for use in this CEL expression: - merge: map1.merge(map2) -> map3 - Merges the passed CEL map with the existing CEL map the function is applied to. - If the same key exists in both maps, if the key's value is type map both maps are merged else the value from the passed map is used. - denormalize: map.denormalize() -> map - Denormalizes a CEL map such that every value of type map or key in the map is expanded to return a single level map. - The resulting keys are \".\" separated indices of the map keys. - For example: { \"a\": 1, \"b\": { \"c\": 2, \"d\": 3 } \"e\": [4, 5] } .denormalize() -> { \"a\": 1, \"b.c\": 2, \"b.d\": 3, \"e.0\": 4, \"e.1\": 5 } - setField: map.setField(key, value) -> message - Sets the field of the message with the given key to the given value. - If the field is not present it will be added. - If the field is present it will be overwritten. - The key can be a dot separated path to set a field in a nested message. - Key must be of type string. - Value may be any valid type. - removeFields: map.removeFields([key1, key2, ...]) -> message - Removes the fields of the map with the given keys. - The keys can be a dot separated path to remove a field in a nested message. - If a key is not found it will be ignored. - Keys must be of type string. - toMap: [map1, map2, ...].toMap() -> map - Converts a CEL list of CEL maps to a single CEL map - toDestinationPayloadFormat(): message.data.toDestinationPayloadFormat() -> string or bytes - Converts the message data to the destination payload format specified in Pipeline.Destination.output_payload_format - This function is meant to be applied to the message.data field. - If the destination payload format is not set, the function will return the message data unchanged. - toCloudEventJsonWithPayloadFormat: message.toCloudEventJsonWithPayloadFormat() -> map - Converts a message to the corresponding structure of JSON format for CloudEvents - This function applies toDestinationPayloadFormat() to the message data. It also sets the corresponding datacontenttype of the CloudEvent, as indicated by Pipeline.Destination.output_payload_format. If no output_payload_format is set it will use the existing datacontenttype on the CloudEvent if present, else leave datacontenttype absent. - This function expects that the content of the message will adhere to the standard CloudEvent format. If it doesn't then this function will fail. - The result is a CEL map that corresponds to the JSON representation of the CloudEvent. To convert that data to a JSON string it can be chained with the toJsonString function.", "type": "string" } }, @@ -3060,12 +3060,12 @@ "type": "integer" }, "maxRetryDelay": { -"description": "Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 60.", +"description": "Optional. The maximum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 60.", "format": "google-duration", "type": "string" }, "minRetryDelay": { -"description": "Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 0 and 600. The default value for this field is 5.", +"description": "Optional. The minimum amount of seconds to wait between retry attempts. The value must be between 1 and 600. The default value for this field is 5.", "format": "google-duration", "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/firebaseappdistribution.v1alpha.json b/googleapiclient/discovery_cache/documents/firebaseappdistribution.v1alpha.json index 86efbdec565..c37ba9efaa7 100644 --- a/googleapiclient/discovery_cache/documents/firebaseappdistribution.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/firebaseappdistribution.v1alpha.json @@ -625,7 +625,7 @@ } } }, -"revision": "20241028", +"revision": "20241101", "rootUrl": "https://firebaseappdistribution.googleapis.com/", "schemas": { "AndroidxCrawlerOutputPoint": { @@ -781,6 +781,14 @@ "goal": { "description": "A goal to be accomplished by the AI", "type": "string" +}, +"hint": { +"description": "Optional. Hint text containing suggestions to help the agent accomplish the goal", +"type": "string" +}, +"successCriteria": { +"description": "Optional. A description of criteria the agent should use to determine if the goal has been successfully completed", +"type": "string" } }, "type": "object" @@ -1489,7 +1497,7 @@ "type": "array" }, "displayName": { -"description": "Optional. Display name of the release test. Required if the release test is created with multiple goals", +"description": "Optional. Display name of the release test. Required if the release test is created with multiple goals.", "type": "string" }, "loginCredential": { @@ -1622,6 +1630,10 @@ "description": "Configuration for automated tests", "id": "GoogleFirebaseAppdistroV1alphaTestConfig", "properties": { +"displayName": { +"description": "Optional. Display name of the AI driven test. Required if the release test is created with multiple goals.", +"type": "string" +}, "name": { "description": "Identifier. The name of the test configuration resource. Format: `projects/{project_number}/apps/{app_id}/testConfig`", "type": "string" diff --git a/googleapiclient/discovery_cache/documents/gkehub.v1.json b/googleapiclient/discovery_cache/documents/gkehub.v1.json index 84506caf0ba..e6c7c737dfa 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v1.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v1.json @@ -2111,7 +2111,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AppDevExperienceFeatureSpec": { @@ -2664,6 +2664,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" @@ -6331,6 +6335,20 @@ "description": "**Service Mesh**: Spec for a single Membership for the servicemesh feature", "id": "ServiceMeshMembershipSpec", "properties": { +"configApi": { +"description": "Optional. Specifies the API that will be used for configuring the mesh workloads.", +"enum": [ +"CONFIG_API_UNSPECIFIED", +"CONFIG_API_ISTIO", +"CONFIG_API_GATEWAY" +], +"enumDescriptions": [ +"Unspecified", +"Use the Istio API for configuration.", +"Use the K8s Gateway API for configuration." +], +"type": "string" +}, "controlPlane": { "deprecated": true, "description": "Deprecated: use `management` instead Enables automatic control plane management.", diff --git a/googleapiclient/discovery_cache/documents/gkehub.v1alpha.json b/googleapiclient/discovery_cache/documents/gkehub.v1alpha.json index e2c21955860..b136fa5029f 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v1alpha.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v1alpha.json @@ -2255,7 +2255,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AnthosObservabilityFeatureSpec": { @@ -3009,6 +3009,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/gkehub.v1beta.json b/googleapiclient/discovery_cache/documents/gkehub.v1beta.json index 4cc5c604ecd..23c49cfb875 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v1beta.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v1beta.json @@ -2111,7 +2111,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AnthosObservabilityFeatureSpec": { @@ -2749,6 +2749,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" @@ -6494,6 +6498,20 @@ "description": "**Service Mesh**: Spec for a single Membership for the servicemesh feature", "id": "ServiceMeshMembershipSpec", "properties": { +"configApi": { +"description": "Optional. Specifies the API that will be used for configuring the mesh workloads.", +"enum": [ +"CONFIG_API_UNSPECIFIED", +"CONFIG_API_ISTIO", +"CONFIG_API_GATEWAY" +], +"enumDescriptions": [ +"Unspecified", +"Use the Istio API for configuration.", +"Use the K8s Gateway API for configuration." +], +"type": "string" +}, "controlPlane": { "deprecated": true, "description": "Deprecated: use `management` instead Enables automatic control plane management.", diff --git a/googleapiclient/discovery_cache/documents/gkehub.v2.json b/googleapiclient/discovery_cache/documents/gkehub.v2.json index 904b1197a34..c1750bfa5d5 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v2.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v2.json @@ -471,7 +471,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AppDevExperienceState": { @@ -729,6 +729,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/gkehub.v2alpha.json b/googleapiclient/discovery_cache/documents/gkehub.v2alpha.json index 414aa5a2bbd..ceed90b499b 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v2alpha.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v2alpha.json @@ -471,7 +471,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AppDevExperienceState": { @@ -729,6 +729,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/gkehub.v2beta.json b/googleapiclient/discovery_cache/documents/gkehub.v2beta.json index 1004226e9c7..94392d4e7f3 100644 --- a/googleapiclient/discovery_cache/documents/gkehub.v2beta.json +++ b/googleapiclient/discovery_cache/documents/gkehub.v2beta.json @@ -471,7 +471,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://gkehub.googleapis.com/", "schemas": { "AppDevExperienceState": { @@ -729,6 +729,10 @@ "sourceFormat": { "description": "Specifies whether the Config Sync Repo is in \"hierarchical\" or \"unstructured\" mode.", "type": "string" +}, +"stopSyncing": { +"description": "Set to true to stop syncing configs for a single cluster. Default to false.", +"type": "boolean" } }, "type": "object" diff --git a/googleapiclient/discovery_cache/documents/gkeonprem.v1.json b/googleapiclient/discovery_cache/documents/gkeonprem.v1.json index 7b512bb3f10..a63cabecd84 100644 --- a/googleapiclient/discovery_cache/documents/gkeonprem.v1.json +++ b/googleapiclient/discovery_cache/documents/gkeonprem.v1.json @@ -1719,6 +1719,49 @@ }, "vmwareAdminClusters": { "methods": { +"create": { +"description": "Creates a new VMware admin cluster in a given project and location. The API needs to be combined with creating a bootstrap cluster to work.", +"flatPath": "v1/projects/{projectsId}/locations/{locationsId}/vmwareAdminClusters", +"httpMethod": "POST", +"id": "gkeonprem.projects.locations.vmwareAdminClusters.create", +"parameterOrder": [ +"parent" +], +"parameters": { +"allowPreflightFailure": { +"description": "Optional. If set to true, CLM will force CCFE to persist the cluster resource in RMS when the creation fails during standalone preflight checks. In that case the subsequent create call will fail with \"cluster already exists\" error and hence a update cluster is required to fix the cluster.", +"location": "query", +"type": "boolean" +}, +"parent": { +"description": "Required. The parent of the project and location where the cluster is created in. Format: \"projects/{project}/locations/{location}\"", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+$", +"required": true, +"type": "string" +}, +"validateOnly": { +"description": "Validate the request without actually doing any updates.", +"location": "query", +"type": "boolean" +}, +"vmwareAdminClusterId": { +"description": "Required. User provided identifier that is used as part of the resource name; must conform to RFC-1034 and additionally restrict to lower-cased letters. This comes out roughly to: /^a-z+[a-z0-9]$/", +"location": "query", +"type": "string" +} +}, +"path": "v1/{+parent}/vmwareAdminClusters", +"request": { +"$ref": "VmwareAdminCluster" +}, +"response": { +"$ref": "Operation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, "enroll": { "description": "Enrolls an existing VMware admin cluster to the Anthos On-Prem API within a given project and location. Through enrollment, an existing admin cluster will become Anthos On-Prem API managed. The corresponding GCP resources will be created and all future modifications to the cluster will be expected to be performed through the API.", "flatPath": "v1/projects/{projectsId}/locations/{locationsId}/vmwareAdminClusters:enroll", @@ -3051,7 +3094,7 @@ } } }, -"revision": "20240814", +"revision": "20241023", "rootUrl": "https://gkeonprem.googleapis.com/", "schemas": { "Authorization": { diff --git a/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json index 45bfd5f6983..e1c48126ef2 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.accounts_v1beta.json @@ -1417,7 +1417,7 @@ } } }, -"revision": "20241028", +"revision": "20241029", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "Accepted": { @@ -1440,7 +1440,7 @@ "type": "object" }, "Account": { -"description": "An account.", +"description": "The `Account` message represents a merchant's account within Shopping Ads. It's the primary entity for managing product data, settings, and interactions with Google's services and external providers. Accounts can operate as standalone entities or be part of a multi-client account (MCA) structure. In an MCA setup the parent account manages multiple sub-accounts. Establishing an account involves configuring attributes like the account name, time zone, and language preferences. The `Account` message is the parent entity for many other resources, for example, `AccountRelationship`, `Homepage`, `BusinessInfo` and so on.", "id": "Account", "properties": { "accountId": { @@ -1671,12 +1671,12 @@ "type": "object" }, "BusinessInfo": { -"description": "Collection of information related to a business.", +"description": "The `BusinessInfo` message contains essential information about a merchant's business. This message captures key business details such as physical address, customer service contacts, and region-specific identifiers.", "id": "BusinessInfo", "properties": { "address": { "$ref": "PostalAddress", -"description": "Optional. The address of the business." +"description": "Optional. The address of the business. Only `region_code`, `address_lines`, `postal_code`, `administrative_area` and `locality` fields are supported. All other fields are ignored." }, "customerService": { "$ref": "CustomerService", @@ -2060,7 +2060,7 @@ "type": "object" }, "Homepage": { -"description": "A store's homepage.", +"description": "The `Homepage` message represents a merchant's store homepage within the system. A merchant's homepage is the primary domain where customers interact with their store. The homepage can be claimed and verified as a proof of ownership and allows the merchant to unlock features that require a verified website. For more information, see [Understanding online store URL verification](//support.google.com/merchants/answer/176793).", "id": "Homepage", "properties": { "claimed": { @@ -3201,7 +3201,7 @@ false "type": "object" }, "TermsOfService": { -"description": "A `TermsOfService`.", +"description": "The `TermsOfService` message represents a specific version of the terms of service that merchants must accept to access certain features or services (see https://support.google.com/merchants/answer/160173). This message is important for the onboarding process, ensuring that merchants agree to the necessary legal agreements for using the service. Merchants can retrieve the latest terms of service for a given `kind` and `region` through `RetrieveLatestTermsOfService`, and accept them as required through `AcceptTermsOfService`.", "id": "TermsOfService", "properties": { "external": { @@ -3351,7 +3351,7 @@ false "type": "object" }, "User": { -"description": "A [user](https://support.google.com/merchants/answer/12160472).", +"description": "The `User` message represents a user associated with a Merchant Center account. It is used to manage user permissions and access rights within the account. For more information, see [Frequently asked questions about people and access levels](//support.google.com/merchants/answer/12160472).", "id": "User", "properties": { "accessRights": { diff --git a/googleapiclient/discovery_cache/documents/merchantapi.conversions_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.conversions_v1beta.json index 7f52ff66b25..63d442943db 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.conversions_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.conversions_v1beta.json @@ -295,7 +295,7 @@ } } }, -"revision": "20241023", +"revision": "20241101", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "AttributionSettings": { @@ -521,7 +521,8 @@ "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -540,6 +541,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -559,7 +561,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/merchantapi.reports_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.reports_v1beta.json index af887e24edf..1d759f48127 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.reports_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.reports_v1beta.json @@ -142,7 +142,7 @@ } } }, -"revision": "20241022", +"revision": "20241101", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "BestSellersBrandView": { @@ -654,7 +654,8 @@ "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -673,6 +674,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -692,7 +694,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "type": "string" } @@ -1031,7 +1034,8 @@ false "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -1050,6 +1054,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -1069,7 +1074,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/merchantapi.reviews_v1beta.json b/googleapiclient/discovery_cache/documents/merchantapi.reviews_v1beta.json index feead767f0e..e8765c40158 100644 --- a/googleapiclient/discovery_cache/documents/merchantapi.reviews_v1beta.json +++ b/googleapiclient/discovery_cache/documents/merchantapi.reviews_v1beta.json @@ -356,7 +356,7 @@ } } }, -"revision": "20241022", +"revision": "20241101", "rootUrl": "https://merchantapi.googleapis.com/", "schemas": { "CustomAttribute": { @@ -569,7 +569,8 @@ "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -588,6 +589,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -607,7 +609,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "readOnly": true, "type": "string" @@ -663,7 +666,8 @@ false "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -682,6 +686,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -701,7 +706,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "readOnly": true, "type": "string" @@ -799,7 +805,8 @@ false "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -818,6 +825,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -837,7 +845,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "type": "string" } @@ -1070,7 +1079,8 @@ false "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -1089,6 +1099,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -1108,7 +1119,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "readOnly": true, "type": "string" @@ -1164,7 +1176,8 @@ false "CLOUD_RETAIL", "LOCAL_CLOUD_RETAIL", "PRODUCT_REVIEWS", -"MERCHANT_REVIEWS" +"MERCHANT_REVIEWS", +"YOUTUBE_CHECKOUT" ], "enumDeprecated": [ false, @@ -1183,6 +1196,7 @@ false, false, false, false, +false, false ], "enumDescriptions": [ @@ -1202,7 +1216,8 @@ false "[Cloud retail](https://cloud.google.com/solutions/retail).", "[Local cloud retail](https://cloud.google.com/solutions/retail).", "[Product Reviews](https://support.google.com/merchants/answer/14620732).", -"[Merchant Reviews](https://developers.google.com/merchant-review-feeds)." +"[Merchant Reviews](https://developers.google.com/merchant-review-feeds).", +"[YouTube Checkout](" ], "readOnly": true, "type": "string" diff --git a/googleapiclient/discovery_cache/documents/migrationcenter.v1alpha1.json b/googleapiclient/discovery_cache/documents/migrationcenter.v1alpha1.json index b09ff778e9b..113a5cadf35 100644 --- a/googleapiclient/discovery_cache/documents/migrationcenter.v1alpha1.json +++ b/googleapiclient/discovery_cache/documents/migrationcenter.v1alpha1.json @@ -2548,7 +2548,7 @@ } } }, -"revision": "20241014", +"revision": "20241027", "rootUrl": "https://migrationcenter.googleapis.com/", "schemas": { "AddAssetsToGroupRequest": { @@ -7818,6 +7818,10 @@ false "description": "Describes the Migration Center settings related to the project.", "id": "Settings", "properties": { +"customerConsentForGoogleSalesToAccessMigrationCenter": { +"description": "Customer consent for Google sales to access their Cloud Migration Center project.", +"type": "boolean" +}, "disableCloudLogging": { "description": "Disable Cloud Logging for the Migration Center API. Users are billed for the logs.", "type": "boolean" diff --git a/googleapiclient/discovery_cache/documents/networkmanagement.v1.json b/googleapiclient/discovery_cache/documents/networkmanagement.v1.json index d59672d3d83..0347247bbc5 100644 --- a/googleapiclient/discovery_cache/documents/networkmanagement.v1.json +++ b/googleapiclient/discovery_cache/documents/networkmanagement.v1.json @@ -591,7 +591,7 @@ } } }, -"revision": "20241016", +"revision": "20241023", "rootUrl": "https://networkmanagement.googleapis.com/", "schemas": { "AbortInfo": { @@ -1340,8 +1340,12 @@ false "readOnly": true, "type": "string" }, +"fqdn": { +"description": "DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address.", +"type": "string" +}, "gkeMasterCluster": { -"description": "A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture).", +"description": "A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture).", "type": "string" }, "instance": { diff --git a/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json b/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json index 950fed5130b..c54b0fd56fc 100644 --- a/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/networkmanagement.v1beta1.json @@ -758,7 +758,7 @@ } } }, -"revision": "20241016", +"revision": "20241023", "rootUrl": "https://networkmanagement.googleapis.com/", "schemas": { "AbortInfo": { @@ -1507,8 +1507,12 @@ false "readOnly": true, "type": "string" }, +"fqdn": { +"description": "DNS endpoint of [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture). Requires gke_master_cluster to be set, can't be used simultaneoulsly with ip_address.", +"type": "string" +}, "gkeMasterCluster": { -"description": "A cluster URI for [Google Kubernetes Engine master](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture).", +"description": "A cluster URI for [Google Kubernetes Engine cluster control plane](https://cloud.google.com/kubernetes-engine/docs/concepts/cluster-architecture).", "type": "string" }, "instance": { diff --git a/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json b/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json index 263d0153697..4ac0fde02d8 100644 --- a/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/networkservices.v1beta1.json @@ -2527,7 +2527,7 @@ } } }, -"revision": "20241010", +"revision": "20241016", "rootUrl": "https://networkservices.googleapis.com/", "schemas": { "AuthzExtension": { @@ -2615,7 +2615,7 @@ ], "enumDescriptions": [ "Not specified.", -"The extension service uses ExtProc GRPC API." +"The extension service uses ExtProc GRPC API over a gRPC stream. This is the default value if the wire format is not specified. The backend service for the extension must use HTTP2 or H2C as the protocol. All `supported_events` for a client request will be sent as part of the same gRPC stream." ], "type": "string" } diff --git a/googleapiclient/discovery_cache/documents/osconfig.v2beta.json b/googleapiclient/discovery_cache/documents/osconfig.v2beta.json index 5c8345bd6c8..64a2a7c8c7a 100644 --- a/googleapiclient/discovery_cache/documents/osconfig.v2beta.json +++ b/googleapiclient/discovery_cache/documents/osconfig.v2beta.json @@ -123,7 +123,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global'", +"description": "Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global`", "location": "path", "pattern": "^folders/[^/]+/locations/global$", "required": true, @@ -440,7 +440,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global'", +"description": "Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global`", "location": "path", "pattern": "^organizations/[^/]+/locations/global$", "required": true, @@ -757,7 +757,7 @@ ], "parameters": { "parent": { -"description": "Required. The parent resource name in the form of: `organizations/{organization_id}/locations/global` `folders/{folder_id}/locations/global` 'projects/{project_id_or_number}/locations/global'", +"description": "Required. The parent resource name in the form of: * `organizations/{organization_id}/locations/global` * `folders/{folder_id}/locations/global` * `projects/{project_id_or_number}/locations/global`", "location": "path", "pattern": "^projects/[^/]+/locations/global$", "required": true, @@ -1057,7 +1057,7 @@ } } }, -"revision": "20241028", +"revision": "20241104", "rootUrl": "https://osconfig.googleapis.com/", "schemas": { "CancelOperationRequest": { diff --git a/googleapiclient/discovery_cache/documents/places.v1.json b/googleapiclient/discovery_cache/documents/places.v1.json index e755ad6204d..e62d04d78b8 100644 --- a/googleapiclient/discovery_cache/documents/places.v1.json +++ b/googleapiclient/discovery_cache/documents/places.v1.json @@ -280,7 +280,7 @@ } } }, -"revision": "20241028", +"revision": "20241029", "rootUrl": "https://places.googleapis.com/", "schemas": { "GoogleGeoTypeViewport": { @@ -1036,7 +1036,6 @@ "type": "array" }, "businessStatus": { -"description": "The business status for the place.", "enum": [ "BUSINESS_STATUS_UNSPECIFIED", "OPERATIONAL", diff --git a/googleapiclient/discovery_cache/documents/recaptchaenterprise.v1.json b/googleapiclient/discovery_cache/documents/recaptchaenterprise.v1.json index 216272da4e7..d5e45119437 100644 --- a/googleapiclient/discovery_cache/documents/recaptchaenterprise.v1.json +++ b/googleapiclient/discovery_cache/documents/recaptchaenterprise.v1.json @@ -786,7 +786,7 @@ } } }, -"revision": "20241013", +"revision": "20241027", "rootUrl": "https://recaptchaenterprise.googleapis.com/", "schemas": { "GoogleCloudRecaptchaenterpriseV1AccountDefenderAssessment": { @@ -1803,7 +1803,7 @@ true "type": "object" }, "GoogleCloudRecaptchaenterpriseV1RemoveIpOverrideRequest": { -"description": "The removeIpOverride request message.", +"description": "The RemoveIpOverride request message.", "id": "GoogleCloudRecaptchaenterpriseV1RemoveIpOverrideRequest", "properties": { "ipOverrideData": { @@ -1855,7 +1855,7 @@ true "id": "GoogleCloudRecaptchaenterpriseV1RiskAnalysis", "properties": { "challenge": { -"description": "Output only. Challenge information for SCORE_AND_CHALLENGE keys", +"description": "Output only. Challenge information for SCORE_AND_CHALLENGE and INVISIBLE keys", "enum": [ "CHALLENGE_UNSPECIFIED", "NOCAPTCHA", diff --git a/googleapiclient/discovery_cache/documents/serviceusage.v1.json b/googleapiclient/discovery_cache/documents/serviceusage.v1.json index 7041614482c..06166225b55 100644 --- a/googleapiclient/discovery_cache/documents/serviceusage.v1.json +++ b/googleapiclient/discovery_cache/documents/serviceusage.v1.json @@ -426,7 +426,7 @@ } } }, -"revision": "20240929", +"revision": "20241025", "rootUrl": "https://serviceusage.googleapis.com/", "schemas": { "AddEnableRulesMetadata": { @@ -2044,6 +2044,171 @@ "properties": {}, "type": "object" }, +"GoogleApiServiceusageV2betaAnalysis": { +"description": "A message to group the analysis information.", +"id": "GoogleApiServiceusageV2betaAnalysis", +"properties": { +"analysis": { +"$ref": "GoogleApiServiceusageV2betaAnalysisResult", +"description": "Output only. Analysis result of updating a policy.", +"readOnly": true +}, +"analysisType": { +"description": "Output only. The type of analysis.", +"enum": [ +"ANALYSIS_TYPE_UNSPECIFIED", +"ANALYSIS_TYPE_DEPENDENCY", +"ANALYSIS_TYPE_RESOURCE_USAGE" +], +"enumDescriptions": [ +"Unspecified analysis type. Do not use.", +"The analysis of service dependencies.", +"The analysis of service resource usage." +], +"readOnly": true, +"type": "string" +}, +"displayName": { +"description": "Output only. The user friendly display name of the analysis type. E.g. service dependency analysis, service resource usage analysis, etc.", +"readOnly": true, +"type": "string" +}, +"service": { +"description": "The names of the service that has analysis result of warnings or blockers. Example: `services/storage.googleapis.com`.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalysisResult": { +"description": "An analysis result including blockers and warnings.", +"id": "GoogleApiServiceusageV2betaAnalysisResult", +"properties": { +"blockers": { +"description": "Blocking information that would prevent the policy changes at runtime.", +"items": { +"$ref": "GoogleApiServiceusageV2betaImpact" +}, +"type": "array" +}, +"warnings": { +"description": "Warning information indicating that the policy changes might be unsafe, but will not block the changes at runtime.", +"items": { +"$ref": "GoogleApiServiceusageV2betaImpact" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalyzeConsumerPolicyMetadata": { +"description": "Metadata for the `AnalyzeConsumerPolicy` method.", +"id": "GoogleApiServiceusageV2betaAnalyzeConsumerPolicyMetadata", +"properties": {}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalyzeConsumerPolicyResponse": { +"description": "The response of analyzing a consumer policy update.", +"id": "GoogleApiServiceusageV2betaAnalyzeConsumerPolicyResponse", +"properties": { +"analysis": { +"description": "The list of analyses returned from performing the intended policy update analysis. The analysis is grouped by service name and different analysis types. The empty analysis list means that the consumer policy can be updated without any warnings or blockers.", +"items": { +"$ref": "GoogleApiServiceusageV2betaAnalysis" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaConsumerPolicy": { +"description": "Consumer Policy is a set of rules that define what services or service groups can be used for a cloud resource hierarchy.", +"id": "GoogleApiServiceusageV2betaConsumerPolicy", +"properties": { +"annotations": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. Annotations is an unstructured key-value map stored with a policy that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. [AIP-128](https://google.aip.dev/128#annotations)", +"type": "object" +}, +"createTime": { +"description": "Output only. The time the policy was created. For singleton policies, this is the first touch of the policy.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"enableRules": { +"description": "Enable rules define usable services, groups, and categories. There can currently be at most one `EnableRule`. This restriction will be lifted in later releases.", +"items": { +"$ref": "GoogleApiServiceusageV2betaEnableRule" +}, +"type": "array" +}, +"etag": { +"description": "Output only. An opaque tag indicating the current version of the policy, used for concurrency control.", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The resource name of the policy. Only the `default` policy is supported: `projects/12345/consumerPolicies/default`, `folders/12345/consumerPolicies/default`, `organizations/12345/consumerPolicies/default`.", +"readOnly": true, +"type": "string" +}, +"updateTime": { +"description": "Output only. The time the policy was last updated.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaEnableRule": { +"description": "The consumer policy rule that defines enabled services, groups, and categories.", +"id": "GoogleApiServiceusageV2betaEnableRule", +"properties": { +"services": { +"description": "The names of the services that are enabled. Example: `services/storage.googleapis.com`.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaImpact": { +"description": "A message to group impacts of updating a policy.", +"id": "GoogleApiServiceusageV2betaImpact", +"properties": { +"detail": { +"description": "Output only. User friendly impact detail in a free form message.", +"readOnly": true, +"type": "string" +}, +"impactType": { +"description": "Output only. The type of impact.", +"enum": [ +"IMPACT_TYPE_UNSPECIFIED", +"DEPENDENCY_MISSING_DEPENDENCIES" +], +"enumDescriptions": [ +"Reserved Blocks (Block n contains codes from 100n to 100(n+1) -1 Block 0 - Special/Admin codes Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY Block 2 - Impact Type of ANALYSIS_TYPE_RESOURCE_USAGE ...", +"Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY" +], +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaUpdateConsumerPolicyMetadata": { +"description": "Metadata for the `UpdateConsumerPolicy` method.", +"id": "GoogleApiServiceusageV2betaUpdateConsumerPolicyMetadata", +"properties": {}, +"type": "object" +}, "Http": { "description": "Defines the HTTP configuration for an API service. It contains a list of HttpRule, each specifying the mapping of an RPC method to one or more HTTP REST API methods.", "id": "Http", diff --git a/googleapiclient/discovery_cache/documents/serviceusage.v1beta1.json b/googleapiclient/discovery_cache/documents/serviceusage.v1beta1.json index fc162c11e0f..f84d3d8834e 100644 --- a/googleapiclient/discovery_cache/documents/serviceusage.v1beta1.json +++ b/googleapiclient/discovery_cache/documents/serviceusage.v1beta1.json @@ -964,7 +964,7 @@ } } }, -"revision": "20240929", +"revision": "20241025", "rootUrl": "https://serviceusage.googleapis.com/", "schemas": { "AddEnableRulesMetadata": { @@ -2621,6 +2621,171 @@ "properties": {}, "type": "object" }, +"GoogleApiServiceusageV2betaAnalysis": { +"description": "A message to group the analysis information.", +"id": "GoogleApiServiceusageV2betaAnalysis", +"properties": { +"analysis": { +"$ref": "GoogleApiServiceusageV2betaAnalysisResult", +"description": "Output only. Analysis result of updating a policy.", +"readOnly": true +}, +"analysisType": { +"description": "Output only. The type of analysis.", +"enum": [ +"ANALYSIS_TYPE_UNSPECIFIED", +"ANALYSIS_TYPE_DEPENDENCY", +"ANALYSIS_TYPE_RESOURCE_USAGE" +], +"enumDescriptions": [ +"Unspecified analysis type. Do not use.", +"The analysis of service dependencies.", +"The analysis of service resource usage." +], +"readOnly": true, +"type": "string" +}, +"displayName": { +"description": "Output only. The user friendly display name of the analysis type. E.g. service dependency analysis, service resource usage analysis, etc.", +"readOnly": true, +"type": "string" +}, +"service": { +"description": "The names of the service that has analysis result of warnings or blockers. Example: `services/storage.googleapis.com`.", +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalysisResult": { +"description": "An analysis result including blockers and warnings.", +"id": "GoogleApiServiceusageV2betaAnalysisResult", +"properties": { +"blockers": { +"description": "Blocking information that would prevent the policy changes at runtime.", +"items": { +"$ref": "GoogleApiServiceusageV2betaImpact" +}, +"type": "array" +}, +"warnings": { +"description": "Warning information indicating that the policy changes might be unsafe, but will not block the changes at runtime.", +"items": { +"$ref": "GoogleApiServiceusageV2betaImpact" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalyzeConsumerPolicyMetadata": { +"description": "Metadata for the `AnalyzeConsumerPolicy` method.", +"id": "GoogleApiServiceusageV2betaAnalyzeConsumerPolicyMetadata", +"properties": {}, +"type": "object" +}, +"GoogleApiServiceusageV2betaAnalyzeConsumerPolicyResponse": { +"description": "The response of analyzing a consumer policy update.", +"id": "GoogleApiServiceusageV2betaAnalyzeConsumerPolicyResponse", +"properties": { +"analysis": { +"description": "The list of analyses returned from performing the intended policy update analysis. The analysis is grouped by service name and different analysis types. The empty analysis list means that the consumer policy can be updated without any warnings or blockers.", +"items": { +"$ref": "GoogleApiServiceusageV2betaAnalysis" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaConsumerPolicy": { +"description": "Consumer Policy is a set of rules that define what services or service groups can be used for a cloud resource hierarchy.", +"id": "GoogleApiServiceusageV2betaConsumerPolicy", +"properties": { +"annotations": { +"additionalProperties": { +"type": "string" +}, +"description": "Optional. Annotations is an unstructured key-value map stored with a policy that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. [AIP-128](https://google.aip.dev/128#annotations)", +"type": "object" +}, +"createTime": { +"description": "Output only. The time the policy was created. For singleton policies, this is the first touch of the policy.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +}, +"enableRules": { +"description": "Enable rules define usable services, groups, and categories. There can currently be at most one `EnableRule`. This restriction will be lifted in later releases.", +"items": { +"$ref": "GoogleApiServiceusageV2betaEnableRule" +}, +"type": "array" +}, +"etag": { +"description": "Output only. An opaque tag indicating the current version of the policy, used for concurrency control.", +"readOnly": true, +"type": "string" +}, +"name": { +"description": "Output only. The resource name of the policy. Only the `default` policy is supported: `projects/12345/consumerPolicies/default`, `folders/12345/consumerPolicies/default`, `organizations/12345/consumerPolicies/default`.", +"readOnly": true, +"type": "string" +}, +"updateTime": { +"description": "Output only. The time the policy was last updated.", +"format": "google-datetime", +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaEnableRule": { +"description": "The consumer policy rule that defines enabled services, groups, and categories.", +"id": "GoogleApiServiceusageV2betaEnableRule", +"properties": { +"services": { +"description": "The names of the services that are enabled. Example: `services/storage.googleapis.com`.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaImpact": { +"description": "A message to group impacts of updating a policy.", +"id": "GoogleApiServiceusageV2betaImpact", +"properties": { +"detail": { +"description": "Output only. User friendly impact detail in a free form message.", +"readOnly": true, +"type": "string" +}, +"impactType": { +"description": "Output only. The type of impact.", +"enum": [ +"IMPACT_TYPE_UNSPECIFIED", +"DEPENDENCY_MISSING_DEPENDENCIES" +], +"enumDescriptions": [ +"Reserved Blocks (Block n contains codes from 100n to 100(n+1) -1 Block 0 - Special/Admin codes Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY Block 2 - Impact Type of ANALYSIS_TYPE_RESOURCE_USAGE ...", +"Block 1 - Impact Type of ANALYSIS_TYPE_DEPENDENCY" +], +"readOnly": true, +"type": "string" +} +}, +"type": "object" +}, +"GoogleApiServiceusageV2betaUpdateConsumerPolicyMetadata": { +"description": "Metadata for the `UpdateConsumerPolicy` method.", +"id": "GoogleApiServiceusageV2betaUpdateConsumerPolicyMetadata", +"properties": {}, +"type": "object" +}, "Http": { "description": "Defines the HTTP configuration for an API service. It contains a list of HttpRule, each specifying the mapping of an RPC method to one or more HTTP REST API methods.", "id": "Http", diff --git a/googleapiclient/discovery_cache/documents/tpu.v2alpha1.json b/googleapiclient/discovery_cache/documents/tpu.v2alpha1.json index a9159f4139d..1738c5d41ba 100644 --- a/googleapiclient/discovery_cache/documents/tpu.v2alpha1.json +++ b/googleapiclient/discovery_cache/documents/tpu.v2alpha1.json @@ -473,6 +473,34 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, +"performMaintenance": { +"description": "Perform manual maintenance on a node.", +"flatPath": "v2alpha1/projects/{projectsId}/locations/{locationsId}/nodes/{nodesId}:performMaintenance", +"httpMethod": "POST", +"id": "tpu.projects.locations.nodes.performMaintenance", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The resource name.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/nodes/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2alpha1/{+name}:performMaintenance", +"request": { +"$ref": "PerformMaintenanceRequest" +}, +"response": { +"$ref": "Operation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, "simulateMaintenanceEvent": { "description": "Simulates a maintenance event.", "flatPath": "v2alpha1/projects/{projectsId}/locations/{locationsId}/nodes/{nodesId}:simulateMaintenanceEvent", @@ -815,6 +843,34 @@ "https://www.googleapis.com/auth/cloud-platform" ] }, +"performMaintenanceQueuedResource": { +"description": "Perform manual maintenance on specific nodes of a QueuedResource.", +"flatPath": "v2alpha1/projects/{projectsId}/locations/{locationsId}/queuedResources/{queuedResourcesId}:performMaintenanceQueuedResource", +"httpMethod": "POST", +"id": "tpu.projects.locations.queuedResources.performMaintenanceQueuedResource", +"parameterOrder": [ +"name" +], +"parameters": { +"name": { +"description": "Required. The name of the QueuedResource which holds the nodes to perform maintenance on.", +"location": "path", +"pattern": "^projects/[^/]+/locations/[^/]+/queuedResources/[^/]+$", +"required": true, +"type": "string" +} +}, +"path": "v2alpha1/{+name}:performMaintenanceQueuedResource", +"request": { +"$ref": "PerformMaintenanceQueuedResourceRequest" +}, +"response": { +"$ref": "Operation" +}, +"scopes": [ +"https://www.googleapis.com/auth/cloud-platform" +] +}, "reset": { "description": "Resets a QueuedResource TPU instance", "flatPath": "v2alpha1/projects/{projectsId}/locations/{locationsId}/queuedResources/{queuedResourcesId}:reset", @@ -965,7 +1021,7 @@ } } }, -"revision": "20241017", +"revision": "20241025", "rootUrl": "https://tpu.googleapis.com/", "schemas": { "AcceleratorConfig": { @@ -984,7 +1040,8 @@ "V3", "V4", "V5LITE_POD", -"V5P" +"V5P", +"V6E" ], "enumDescriptions": [ "Unspecified version.", @@ -992,7 +1049,8 @@ "TPU v3.", "TPU v4.", "TPU v5lite pod.", -"TPU v5." +"TPU v5.", +"TPU v6e." ], "type": "string" } @@ -1733,6 +1791,11 @@ "type": "string" }, "type": "array" +}, +"upcomingMaintenance": { +"$ref": "UpcomingMaintenance", +"description": "Output only. Upcoming maintenance on this TPU node.", +"readOnly": true } }, "type": "object" @@ -1832,6 +1895,26 @@ }, "type": "object" }, +"PerformMaintenanceQueuedResourceRequest": { +"description": "Request for PerformMaintenanceQueuedResource.", +"id": "PerformMaintenanceQueuedResourceRequest", +"properties": { +"nodeNames": { +"description": "The names of the nodes to perform maintenance on.", +"items": { +"type": "string" +}, +"type": "array" +} +}, +"type": "object" +}, +"PerformMaintenanceRequest": { +"description": "Request for PerformMaintenance.", +"id": "PerformMaintenanceRequest", +"properties": {}, +"type": "object" +}, "ProvisioningData": { "description": "Further data for the provisioning state.", "id": "ProvisioningData", @@ -2277,6 +2360,57 @@ }, "type": "object" }, +"UpcomingMaintenance": { +"description": "Upcoming Maintenance notification information.", +"id": "UpcomingMaintenance", +"properties": { +"canReschedule": { +"description": "Indicates if the maintenance can be customer triggered.", +"type": "boolean" +}, +"latestWindowStartTime": { +"description": "The latest time for the planned maintenance window to start. This timestamp value is in RFC3339 text format.", +"type": "string" +}, +"maintenanceStatus": { +"description": "The status of the maintenance.", +"enum": [ +"UNKNOWN", +"PENDING", +"ONGOING" +], +"enumDescriptions": [ +"Unknown maintenance status. Do not use this value.", +"There is pending maintenance.", +"There is ongoing maintenance on this VM." +], +"type": "string" +}, +"type": { +"description": "Defines the type of maintenance.", +"enum": [ +"UNKNOWN_TYPE", +"SCHEDULED", +"UNSCHEDULED" +], +"enumDescriptions": [ +"No type specified. Do not use this value.", +"Scheduled maintenance (e.g. maintenance after uptime guarantee is complete).", +"Unscheduled maintenance (e.g. emergency maintenance during uptime guarantee)." +], +"type": "string" +}, +"windowEndTime": { +"description": "The time by which the maintenance disruption will be completed. This timestamp value is in RFC3339 text format.", +"type": "string" +}, +"windowStartTime": { +"description": "The current start time of the maintenance window. This timestamp value is in RFC3339 text format.", +"type": "string" +} +}, +"type": "object" +}, "Usage": { "id": "Usage", "properties": { diff --git a/googleapiclient/discovery_cache/documents/translate.v3.json b/googleapiclient/discovery_cache/documents/translate.v3.json index 13eb0b1eca9..1508c35e287 100644 --- a/googleapiclient/discovery_cache/documents/translate.v3.json +++ b/googleapiclient/discovery_cache/documents/translate.v3.json @@ -291,7 +291,8 @@ "$ref": "Operation" }, "scopes": [ -"https://www.googleapis.com/auth/cloud-platform" +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-translation" ] }, "batchTranslateText": { @@ -319,7 +320,8 @@ "$ref": "Operation" }, "scopes": [ -"https://www.googleapis.com/auth/cloud-platform" +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-translation" ] }, "detectLanguage": { @@ -1708,7 +1710,7 @@ } } }, -"revision": "20240912", +"revision": "20241029", "rootUrl": "https://translation.googleapis.com/", "schemas": { "AdaptiveMtDataset": { diff --git a/googleapiclient/discovery_cache/documents/translate.v3beta1.json b/googleapiclient/discovery_cache/documents/translate.v3beta1.json index 98145fc3d4c..5f82c0f7cc3 100644 --- a/googleapiclient/discovery_cache/documents/translate.v3beta1.json +++ b/googleapiclient/discovery_cache/documents/translate.v3beta1.json @@ -233,7 +233,8 @@ "$ref": "Operation" }, "scopes": [ -"https://www.googleapis.com/auth/cloud-platform" +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-translation" ] }, "batchTranslateText": { @@ -261,7 +262,8 @@ "$ref": "Operation" }, "scopes": [ -"https://www.googleapis.com/auth/cloud-platform" +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-translation" ] }, "detectLanguage": { @@ -484,7 +486,8 @@ "$ref": "Operation" }, "scopes": [ -"https://www.googleapis.com/auth/cloud-platform" +"https://www.googleapis.com/auth/cloud-platform", +"https://www.googleapis.com/auth/cloud-translation" ] }, "delete": { @@ -744,7 +747,7 @@ } } }, -"revision": "20240301", +"revision": "20241029", "rootUrl": "https://translation.googleapis.com/", "schemas": { "BatchDocumentInputConfig": { diff --git a/googleapiclient/discovery_cache/documents/youtube.v3.json b/googleapiclient/discovery_cache/documents/youtube.v3.json index e4d36e3a61c..1afa12af7b9 100644 --- a/googleapiclient/discovery_cache/documents/youtube.v3.json +++ b/googleapiclient/discovery_cache/documents/youtube.v3.json @@ -4133,7 +4133,7 @@ } } }, -"revision": "20241022", +"revision": "20241031", "rootUrl": "https://youtube.googleapis.com/", "schemas": { "AbuseReport": { @@ -10076,7 +10076,7 @@ true }, "kind": { "default": "youtube#playlistItemListResponse", -"description": "Identifies what kind of resource this is. Value: the fixed string \"youtube#playlistItemListResponse\". Etag of this resource.", +"description": "Identifies what kind of resource this is. Value: the fixed string \"youtube#playlistItemListResponse\".", "type": "string" }, "nextPageToken": { @@ -12018,6 +12018,10 @@ true "description": "Basic details about a video category, such as its localized title. Next Id: 19", "id": "VideoStatus", "properties": { +"containsSyntheticMedia": { +"description": "Indicates if the video contains altered or synthetic media.", +"type": "boolean" +}, "embeddable": { "description": "This value indicates if the video can be embedded on another website. @mutable youtube.videos.insert youtube.videos.update", "type": "boolean"