diff --git a/CHANGELOG.md b/CHANGELOG.md index 74fff96275a..82642fddf52 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,48 @@ # Changelog +## [2.132.0](https://github.com/googleapis/google-api-python-client/compare/v2.131.0...v2.132.0) (2024-06-04) + + +### Features + +* **aiplatform:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/9d6000fa065ac1ef877de37b94a5e923c89b8228 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **analyticsadmin:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/494a29d2266725566e185c41e19c08419c88f9b4 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **androidmanagement:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/5afc4010f2f7d303ba0b3a812aab7496aea97adb ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **backupdr:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/5bcc5d39d04aa4691e36cc57b256d983ec52159b ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **chromemanagement:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/32ddf526ff40d30f20f9116027a4f208f38cc792 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **cloudbilling:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/2b5c66b2c5d2ffaa649dd9455da765e10dbce113 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **cloudfunctions:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/34314fb79a2ef113f2f1db15738f2d2e29887222 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **cloudsearch:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/d32e900aeae99a2d7cab64037a2a0d8285aba8b6 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **compute:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/4f7da21c3c67d1019b996492e5dfc9dcacb38214 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **connectors:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/8087f14f8942261881ea87bf47fba512a78a9fc1 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **contactcenteraiplatform:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/4fb577d2d6e2851c8d923066c9ff7b5c1e9df79e ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **contactcenterinsights:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/bb49784a9cb793ff64c8e1d4ee3b98a173b4e31d ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **datamigration:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/ac474a90aeb6d2443b12c1bf891c7fb81dbcb9ed ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **dataplex:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/d959b3d78c7034bbc3571d9ede7d6de3587989f7 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **datastream:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/3abd0f41f2e617749aba78913cb4fa6391df55a8 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **dialogflow:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/2d79840e8bfc7aa3bee79b9554627dfd1cb13121 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **discoveryengine:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/4522cd5e31c6437d52d8d8a09a54cf2c38fb7dcf ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **documentai:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/a06827efcc41fe6af56f687f7c1dc4f8538a166b ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **fcmdata:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/f7c50fd9f7b75df93ef9775684cba47b66cb0c81 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **firebaseappcheck:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/0744228b03e4c38e64358d9b38c17b2df3e2871e ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **healthcare:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/05c4657fa6322067b421e9e0d887904faba04811 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **iam:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/331029f3a230aa25f32a75b9e81adf9d6ed97ed5 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **integrations:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/8bd4954709fc4bea245abd2efca870e8fdbc2c40 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **migrationcenter:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/b46b8b7081691a40f80241bfa154acc6d46abc9d ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **networkconnectivity:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/ff49e0b244002d44580f689e0a3f77175bbe5dfb ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **policyanalyzer:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/b56b2b1453126a06a9bcba1c96766a905006d3a7 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **resourcesettings:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/a5e25b381450da4c88bf86d24550fa7a75f4636a ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **run:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/81892c895bfe7d8b5a60a1ce7c62f6bbd603a7b0 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **servicecontrol:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/0cfcab3609ec38a84d245cc3207cedc6ec92db5a ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **spanner:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/195cae366ac9c01537584735879ef5ae658efee2 ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **versionhistory:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/9cef71c5a52655e5e37b51ac0a430801c2cd97bd ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) +* **workflowexecutions:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/6670b1ea9d65e7574d77954cfd1722736bfa5d1c ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) + + +### Bug Fixes + +* **secretmanager:** Update the api https://togithub.com/googleapis/google-api-python-client/commit/d0199eaf1f51289ad13683a54b6b26a5019b560d ([9401d1f](https://github.com/googleapis/google-api-python-client/commit/9401d1f1f16ed979d217e540aa044f430699aa4d)) + ## [2.131.0](https://github.com/googleapis/google-api-python-client/compare/v2.130.0...v2.131.0) (2024-05-28) diff --git a/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html b/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html index 312da3ee01a..41141273d66 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.batchPredictionJobs.html @@ -360,7 +360,7 @@
Retrieves the next page of results.
+
+ patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a NotebookRuntimeTemplate.
setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
@@ -124,6 +127,9 @@patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates a NotebookRuntimeTemplate. + +Args: + name: string, The resource name of the NotebookRuntimeTemplate. (required) + body: object, The request body. + The object takes the form of: + +{ # A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template. + "createTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was created. + "dataPersistentDiskSpec": { # Represents the spec of persistent disk options. # Optional. The specification of persistent disk attached to the runtime as data disk storage. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + "description": "A String", # The description of the NotebookRuntimeTemplate. + "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, + "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. + "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. + "eucDisabled": True or False, # Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate. + }, + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, + "isDefault": True or False, # Output only. The default template to use if not specified. + "labels": { # The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. + "a_key": "A String", + }, + "machineSpec": { # Specification of a single machine. # Optional. Immutable. The specification of a single machine for the template. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "name": "A String", # The resource name of the NotebookRuntimeTemplate. + "networkSpec": { # Network spec. # Optional. Network spec. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "networkTags": [ # Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). + "A String", + ], + "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. + "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. + "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. + "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. + }, + "updateTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated. +} + + updateMask: string, Required. The update mask applies to the resource. For the `FieldMask` definition, see google.protobuf.FieldMask. Input format: `{paths: "${updated_filed}"}` Updatable fields: * `encryption_spec.kms_key_name` + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template. + "createTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was created. + "dataPersistentDiskSpec": { # Represents the spec of persistent disk options. # Optional. The specification of persistent disk attached to the runtime as data disk storage. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + "description": "A String", # The description of the NotebookRuntimeTemplate. + "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, + "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. + "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. + "eucDisabled": True or False, # Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate. + }, + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, + "isDefault": True or False, # Output only. The default template to use if not specified. + "labels": { # The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. + "a_key": "A String", + }, + "machineSpec": { # Specification of a single machine. # Optional. Immutable. The specification of a single machine for the template. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "name": "A String", # The resource name of the NotebookRuntimeTemplate. + "networkSpec": { # Network spec. # Optional. Network spec. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "networkTags": [ # Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). + "A String", + ], + "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. + "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. + "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. + "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. + }, + "updateTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated. +}+
setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors. diff --git a/docs/dyn/aiplatform_v1.projects.locations.notebookRuntimes.html b/docs/dyn/aiplatform_v1.projects.locations.notebookRuntimes.html index ad194aae5d3..8c03d6565fe 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.notebookRuntimes.html +++ b/docs/dyn/aiplatform_v1.projects.locations.notebookRuntimes.html @@ -113,8 +113,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -128,13 +135,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -234,8 +234,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -249,13 +256,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -292,8 +292,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -307,13 +314,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. diff --git a/docs/dyn/aiplatform_v1.projects.locations.persistentResources.html b/docs/dyn/aiplatform_v1.projects.locations.persistentResources.html index 6af3e0a394d..27a0ec6128c 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.persistentResources.html +++ b/docs/dyn/aiplatform_v1.projects.locations.persistentResources.html @@ -163,9 +163,20 @@Method Details
}, ], "resourceRuntime": { # Persistent Cluster runtime information as output # Output only. Runtime information of the Persistent Resource. + "accessUris": { # Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" } + "a_key": "A String", + }, }, "resourceRuntimeSpec": { # Configuration for the runtime on a PersistentResource instance, including but not limited to: * Service accounts used to run the workloads. * Whether to make it a dedicated Ray Cluster. # Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration. "raySpec": { # Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes. # Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource. + "headNodeResourcePoolId": "A String", # Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set. + "imageUri": "A String", # Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from [Vertex prebuilt images](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field. + "rayMetricSpec": { # Configuration for the Ray metrics. # Optional. Ray metrics configurations. + "disabled": True or False, # Optional. Flag to disable the Ray metrics collection. + }, + "resourcePoolImages": { # Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" } + "a_key": "A String", + }, }, "serviceAccountSpec": { # Configuration for the use of custom service account to run the workloads. # Optional. Configure the use of workload identity on the PersistentResource "enableCustomServiceAccount": True or False, # Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents). @@ -301,9 +312,20 @@Method Details
}, ], "resourceRuntime": { # Persistent Cluster runtime information as output # Output only. Runtime information of the Persistent Resource. + "accessUris": { # Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" } + "a_key": "A String", + }, }, "resourceRuntimeSpec": { # Configuration for the runtime on a PersistentResource instance, including but not limited to: * Service accounts used to run the workloads. * Whether to make it a dedicated Ray Cluster. # Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration. "raySpec": { # Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes. # Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource. + "headNodeResourcePoolId": "A String", # Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set. + "imageUri": "A String", # Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from [Vertex prebuilt images](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field. + "rayMetricSpec": { # Configuration for the Ray metrics. # Optional. Ray metrics configurations. + "disabled": True or False, # Optional. Flag to disable the Ray metrics collection. + }, + "resourcePoolImages": { # Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" } + "a_key": "A String", + }, }, "serviceAccountSpec": { # Configuration for the use of custom service account to run the workloads. # Optional. Configure the use of workload identity on the PersistentResource "enableCustomServiceAccount": True or False, # Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents). @@ -380,9 +402,20 @@Method Details
}, ], "resourceRuntime": { # Persistent Cluster runtime information as output # Output only. Runtime information of the Persistent Resource. + "accessUris": { # Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" } + "a_key": "A String", + }, }, "resourceRuntimeSpec": { # Configuration for the runtime on a PersistentResource instance, including but not limited to: * Service accounts used to run the workloads. * Whether to make it a dedicated Ray Cluster. # Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration. "raySpec": { # Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes. # Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource. + "headNodeResourcePoolId": "A String", # Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set. + "imageUri": "A String", # Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from [Vertex prebuilt images](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field. + "rayMetricSpec": { # Configuration for the Ray metrics. # Optional. Ray metrics configurations. + "disabled": True or False, # Optional. Flag to disable the Ray metrics collection. + }, + "resourcePoolImages": { # Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" } + "a_key": "A String", + }, }, "serviceAccountSpec": { # Configuration for the use of custom service account to run the workloads. # Optional. Configure the use of workload identity on the PersistentResource "enableCustomServiceAccount": True or False, # Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents). @@ -465,9 +498,20 @@Method Details
}, ], "resourceRuntime": { # Persistent Cluster runtime information as output # Output only. Runtime information of the Persistent Resource. + "accessUris": { # Output only. URIs for user to connect to the Cluster. Example: { "RAY_HEAD_NODE_INTERNAL_IP": "head-node-IP:10001" "RAY_DASHBOARD_URI": "ray-dashboard-address:8888" } + "a_key": "A String", + }, }, "resourceRuntimeSpec": { # Configuration for the runtime on a PersistentResource instance, including but not limited to: * Service accounts used to run the workloads. * Whether to make it a dedicated Ray Cluster. # Optional. Persistent Resource runtime spec. For example, used for Ray cluster configuration. "raySpec": { # Configuration information for the Ray cluster. For experimental launch, Ray cluster creation and Persistent cluster creation are 1:1 mapping: We will provision all the nodes within the Persistent cluster as Ray nodes. # Optional. Ray cluster configuration. Required when creating a dedicated RayCluster on the PersistentResource. + "headNodeResourcePoolId": "A String", # Optional. This will be used to indicate which resource pool will serve as the Ray head node(the first node within that pool). Will use the machine from the first workerpool as the head node by default if this field isn't set. + "imageUri": "A String", # Optional. Default image for user to choose a preferred ML framework (for example, TensorFlow or Pytorch) by choosing from [Vertex prebuilt images](https://cloud.google.com/vertex-ai/docs/training/pre-built-containers). Either this or the resource_pool_images is required. Use this field if you need all the resource pools to have the same Ray image. Otherwise, use the {@code resource_pool_images} field. + "rayMetricSpec": { # Configuration for the Ray metrics. # Optional. Ray metrics configurations. + "disabled": True or False, # Optional. Flag to disable the Ray metrics collection. + }, + "resourcePoolImages": { # Optional. Required if image_uri isn't set. A map of resource_pool_id to prebuild Ray image if user need to use different images for different head/worker pools. This map needs to cover all the resource pool ids. Example: { "ray_head_node_pool": "head image" "ray_worker_node_pool1": "worker image" "ray_worker_node_pool2": "another worker image" } + "a_key": "A String", + }, }, "serviceAccountSpec": { # Configuration for the use of custom service account to run the workloads. # Optional. Configure the use of workload identity on the PersistentResource "enableCustomServiceAccount": True or False, # Required. If true, custom user-managed service account is enforced to run any workloads (for example, Vertex Jobs) on the resource. Otherwise, uses the [Vertex AI Custom Code Service Agent](https://cloud.google.com/vertex-ai/docs/general/access-control#service-agents). diff --git a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html index 4274fe9f508..4a50a55476b 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1.projects.locations.publishers.models.html @@ -258,7 +258,34 @@Method Details
"maxOutputTokens": 42, # Optional. The maximum number of output tokens to generate per message. "presencePenalty": 3.14, # Optional. Positive penalties. "responseMimeType": "A String", # Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. - "responseStyle": "A String", # Optional. Control Three levels of creativity in the model output. Default: RESPONSE_STYLE_BALANCED + "responseSchema": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. The `Schema` object allows the definition of input and output data types. These types can be objects, but also primitives and arrays. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). If set, a compatible response_mime_type must also be set. Compatible mimetypes: `application/json`: Schema for JSON response. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, "stopSequences": [ # Optional. Stop sequences. "A String", ], @@ -305,6 +332,14 @@Method Details
], "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Tool config. This config is shared for all tools provided in the request. + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, "tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. @@ -341,6 +376,8 @@Method Details
}, }, ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. @@ -770,7 +807,34 @@Method Details
"maxOutputTokens": 42, # Optional. The maximum number of output tokens to generate per message. "presencePenalty": 3.14, # Optional. Positive penalties. "responseMimeType": "A String", # Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. - "responseStyle": "A String", # Optional. Control Three levels of creativity in the model output. Default: RESPONSE_STYLE_BALANCED + "responseSchema": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. The `Schema` object allows the definition of input and output data types. These types can be objects, but also primitives and arrays. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). If set, a compatible response_mime_type must also be set. Compatible mimetypes: `application/json`: Schema for JSON response. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, "stopSequences": [ # Optional. Stop sequences. "A String", ], @@ -817,6 +881,14 @@Method Details
], "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Tool config. This config is shared for all tools provided in the request. + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, "tools": [ # Optional. A list of `Tools` the model may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. @@ -853,6 +925,8 @@Method Details
}, }, ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. diff --git a/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html b/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html index 37145703de7..312429ac345 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html +++ b/docs/dyn/aiplatform_v1.projects.locations.trainingPipelines.html @@ -227,7 +227,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -245,7 +245,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -536,7 +536,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -554,7 +554,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -887,7 +887,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -905,7 +905,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1210,7 +1210,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1228,7 +1228,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], diff --git a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html index 34bb882cdff..84551776481 100644 --- a/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1.projects.locations.tuningJobs.html @@ -175,6 +175,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -314,6 +582,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -460,6 +996,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -612,6 +1416,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. diff --git a/docs/dyn/aiplatform_v1.publishers.models.html b/docs/dyn/aiplatform_v1.publishers.models.html index 083fd0a9f82..5a326d1bac6 100644 --- a/docs/dyn/aiplatform_v1.publishers.models.html +++ b/docs/dyn/aiplatform_v1.publishers.models.html @@ -161,7 +161,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -179,7 +179,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html index 49a2e930ee5..07337c58301 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.batchPredictionJobs.html @@ -490,7 +490,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -508,7 +508,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -890,7 +890,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -908,7 +908,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1332,7 +1332,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1350,7 +1350,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1745,7 +1745,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1763,7 +1763,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html b/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html new file mode 100644 index 00000000000..ee23f781c9c --- /dev/null +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.cachedContents.html @@ -0,0 +1,1259 @@ + + + +Vertex AI API . projects . locations . cachedContents
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, x__xgafv=None)
Creates cached content, this call will initialize the cached content in the data storage, and users need to pay for the cache data storage.
+ +Deletes cached content
+ +Gets cached content configurations
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists cached contents in a project
+ +Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates cached content configurations
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, x__xgafv=None)
+Creates cached content, this call will initialize the cached content in the data storage, and users need to pay for the cache data storage. + +Args: + parent: string, Required. The parent resource where the cached content will be created (required) + body: object, The request body. + The object takes the form of: + +{ # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. +}+++ +delete(name, x__xgafv=None)
+Deletes cached content + +Args: + name: string, Required. The resource name referring to the cached content (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+++ +get(name, x__xgafv=None)
+Gets cached content configurations + +Args: + name: string, Required. The resource name referring to the cached content (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists cached contents in a project + +Args: + parent: string, Required. The parent, which owns this collection of cached contents. (required) + pageSize: integer, Optional. The maximum number of cached contents to return. The service may return fewer than this value. If unspecified, some default (under maximum) number of items will be returned. The maximum value is 1000; values above 1000 will be coerced to 1000. + pageToken: string, Optional. A page token, received from a previous `ListCachedContents` call. Provide this to retrieve the subsequent page. When paginating, all other parameters provided to `ListCachedContents` must match the call that provided the page token. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response with a list of CachedContents. + "cachedContents": [ # List of cached contents. + { # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. + }, + ], + "nextPageToken": "A String", # A token, which can be sent as `page_token` to retrieve the next page. If this field is omitted, there are no subsequent pages. +}+++ +list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html index 54ba0cd581b..66609557ec4 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.deploymentResourcePools.html @@ -137,7 +137,12 @@patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates cached content configurations + +Args: + name: string, Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} (required) + body: object, The request body. + The object takes the form of: + +{ # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. +} + + updateMask: string, Required. The list of fields to update. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A resource used in LLM queries for users to explicitly specify what to cache and how to cache. + "contents": [ # Optional. Input only. Immutable. The content to cache + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "createTime": "A String", # Output only. Creatation time of the cache entry. + "expireTime": "A String", # Timestamp of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input. + "model": "A String", # Immutable. The name of the publisher model to use for cached content. Format: projects/{project}/locations/{location}/publishers/{publisher}/models/{model} + "name": "A String", # Immutable. Identifier. The resource name of the cached content Format: projects/{project}/locations/{location}/cachedContents/{cached_content} + "systemInstruction": { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. # Optional. Input only. Immutable. Developer set system instruction. Currently, text only + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + "toolConfig": { # Tool config. This config is shared for all tools provided in the request. # Optional. Input only. Immutable. Tool config. This config is shared for all tools + "functionCallingConfig": { # Function calling config. # Optional. Function calling config. + "allowedFunctionNames": [ # Optional. Function names to call. Only set when the Mode is ANY. Function names should match [FunctionDeclaration.name]. With mode set to ANY, model will predict a function call from the set of function names provided. + "A String", + ], + "mode": "A String", # Optional. Function calling mode. + }, + }, + "tools": [ # Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response + { # Tool details that the model may use to generate response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the model. A Tool object should contain exactly one type of Tool (e.g FunctionDeclaration, Retrieval or GoogleSearchRetrieval). + "functionDeclarations": [ # Optional. Function tool type. One or more function declarations to be passed to the model along with the current user query. Model may decide to call a subset of these functions by populating FunctionCall in the response. User should provide a FunctionResponse for each function call in the next turn. Based on the function responses, Model will generate the final response back to the user. Maximum 64 function declarations can be provided. + { # Structured representation of a function declaration as defined by the [OpenAPI 3.0 specification](https://spec.openapis.org/oas/v3.0.3). Included in this declaration are the function name and parameters. This FunctionDeclaration is a representation of a block of code that can be used as a `Tool` by the model and executed by the client. + "description": "A String", # Optional. Description and purpose of the function. Model uses it to decide how and whether to call the function. + "name": "A String", # Required. The name of the function to call. Must start with a letter or an underscore. Must be a-z, A-Z, 0-9, or contain underscores, dots and dashes, with a maximum length of 64. + "parameters": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the parameters to this function in JSON Schema Object format. Reflects the Open API 3.03 Parameter Object. string Key: the name of the parameter. Parameter names are case sensitive. Schema Value: the Schema defining the type used for the parameter. For function with no parameters, this can be left unset. Parameter names must start with a letter or an underscore and must only contain chars a-z, A-Z, 0-9, or underscores with a maximum length of 64. Example with 1 required and 1 optional parameter: type: OBJECT properties: param1: type: STRING param2: type: INTEGER required: - param1 + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + "response": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. Describes the output from this function in JSON Schema format. Reflects the Open API 3.03 Response Object. The Schema defines the type used for the response value of the function. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, + }, + ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, + "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. + "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. + "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. + "datastore": "A String", # Required. Fully-qualified Vertex AI Search's datastore resource ID. Format: `projects/{project}/locations/{location}/collections/{collection}/dataStores/{dataStore}` + }, + "vertexRagStore": { # Retrieve from Vertex RAG Store for grounding. # Set to use data source powered by Vertex RAG store. User data is uploaded via the VertexRagDataService. + "ragCorpora": [ # Optional. Deprecated. Please use rag_resources instead. + "A String", + ], + "ragResources": [ # Optional. The representation of the rag source. It can be used to specify corpus only or ragfiles. Currently only support one corpus or multiple files from one corpus. In the future we may open up multiple corpora support. + { # The definition of the Rag resource. + "ragCorpus": "A String", # Optional. RagCorpora resource name. Format: `projects/{project}/locations/{location}/ragCorpora/{rag_corpus}` + "ragFileIds": [ # Optional. rag_file_id. The files should be in the same rag_corpus set in rag_corpus field. + "A String", + ], + }, + ], + "similarityTopK": 42, # Optional. Number of top k results to return from the selected corpora. + "vectorDistanceThreshold": 3.14, # Optional. Only return results with vector distance smaller than the threshold. + }, + }, + }, + ], + "ttl": "A String", # Input only. The TTL for this resource. The expiration time is computed: now + TTL. + "updateTime": "A String", # Output only. When the cache entry was last updated in UTC time. +}+Method Details
"maxReplicaCount": 42, # Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). "minReplicaCount": 42, # Required. Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed. }, + "disableContainerLogging": True or False, # If the DeploymentResourcePool is deployed with custom-trained Models or AutoML Tabular Models, the container(s) of the DeploymentResourcePool will send `stderr` and `stdout` streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging/pricing). User can disable container logging by setting this flag to true. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for a DeploymentResourcePool. If set, this DeploymentResourcePool will be secured by this key. Endpoints and the DeploymentResourcePool they deploy in need to have the same EncryptionSpec. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "name": "A String", # Immutable. The resource name of the DeploymentResourcePool. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}` + "serviceAccount": "A String", # The service account that the DeploymentResourcePool's container(s) run as. Specify the email address of the service account. If this service account is not specified, the container(s) run as a service account that doesn't have access to the resource project. Users deploying the Models to this DeploymentResourcePool must have the `iam.serviceAccounts.actAs` permission on this service account. }, "deploymentResourcePoolId": "A String", # Required. The ID to use for the DeploymentResourcePool, which will become the final component of the DeploymentResourcePool's resource name. The maximum length is 63 characters, and valid characters are `/^[a-z]([a-z0-9-]{0,61}[a-z0-9])?$/`. } @@ -238,7 +243,12 @@Method Details
"maxReplicaCount": 42, # Immutable. The maximum number of replicas this DeployedModel may be deployed on when the traffic against it increases. If the requested value is too large, the deployment will error, but if deployment succeeds then the ability to scale the model to that many replicas is guaranteed (barring service outages). If traffic against the DeployedModel increases beyond what its replicas at maximum may handle, a portion of the traffic will be dropped. If this value is not provided, will use min_replica_count as the default value. The value of this field impacts the charge against Vertex CPU and GPU quotas. Specifically, you will be charged for (max_replica_count * number of cores in the selected machine type) and (max_replica_count * number of GPUs per replica in the selected machine type). "minReplicaCount": 42, # Required. Immutable. The minimum number of machine replicas this DeployedModel will be always deployed on. This value must be greater than or equal to 1. If traffic against the DeployedModel increases, it may dynamically be deployed onto more replicas, and as traffic decreases, some of these extra replicas may be freed. }, + "disableContainerLogging": True or False, # If the DeploymentResourcePool is deployed with custom-trained Models or AutoML Tabular Models, the container(s) of the DeploymentResourcePool will send `stderr` and `stdout` streams to Cloud Logging by default. Please note that the logs incur cost, which are subject to [Cloud Logging pricing](https://cloud.google.com/logging/pricing). User can disable container logging by setting this flag to true. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for a DeploymentResourcePool. If set, this DeploymentResourcePool will be secured by this key. Endpoints and the DeploymentResourcePool they deploy in need to have the same EncryptionSpec. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "name": "A String", # Immutable. The resource name of the DeploymentResourcePool. Format: `projects/{project}/locations/{location}/deploymentResourcePools/{deployment_resource_pool}` + "serviceAccount": "A String", # The service account that the DeploymentResourcePool's container(s) run as. Specify the email address of the service account. If this service account is not specified, the container(s) run as a service account that doesn't have access to the resource project. Users deploying the Models to this DeploymentResourcePool must have the `iam.serviceAccounts.actAs` permission on this service account. }
+ chat()
+
Returns the chat Resource.
+ @@ -1209,6 +1214,7 @@Returns the batchPredictionJobs Resource.
+
+ cachedContents()
+
Returns the cachedContents Resource.
+ diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.metadataStores.html b/docs/dyn/aiplatform_v1beta1.projects.locations.metadataStores.html index 0ec4bc5c407..6667a821eca 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.metadataStores.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.metadataStores.html @@ -134,6 +134,9 @@Close httplib2 connections.
+
+ create(parent, body=None, notebookExecutionJobId=None, x__xgafv=None)
Creates a NotebookExecutionJob.
Deletes a NotebookExecutionJob.
@@ -101,6 +104,82 @@Close httplib2 connections.+
create(parent, body=None, notebookExecutionJobId=None, x__xgafv=None)
+ Creates a NotebookExecutionJob. + +Args: + parent: string, Required. The resource name of the Location to create the NotebookExecutionJob. Format: `projects/{project}/locations/{location}` (required) + body: object, The request body. + The object takes the form of: + +{ # NotebookExecutionJob represents an instance of a notebook execution. + "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. + "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. + "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. + "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` + }, + "directNotebookSource": { # The content of the input notebook in ipynb format. # The contents of an input notebook file. + "content": "A String", # The base64-encoded contents of the input notebook file. + }, + "displayName": "A String", # The display name of the NotebookExecutionJob. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "executionTimeout": "A String", # Max running time of the execution job in seconds (default 86400s / 24 hrs). + "executionUser": "A String", # The user email to run the execution as. Only supported by Colab runtimes. + "gcsNotebookSource": { # The Cloud Storage uri for the input notebook. # The Cloud Storage url pointing to the ipynb file. Format: `gs://bucket/notebook_file.ipynb` + "generation": "A String", # The version of the Cloud Storage object to read. If unset, the current version of the object is read. See https://cloud.google.com/storage/docs/metadata#generation-number. + "uri": "A String", # The Cloud Storage uri pointing to the ipynb file. Format: `gs://bucket/notebook_file.ipynb` + }, + "gcsOutputUri": "A String", # The Cloud Storage location to upload the result to. Format: `gs://bucket-name` + "jobState": "A String", # Output only. The state of the NotebookExecutionJob. + "labels": { # The labels with user-defined metadata to organize NotebookExecutionJobs. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. + "a_key": "A String", + }, + "name": "A String", # Output only. The resource name of this NotebookExecutionJob. Format: `projects/{project_id}/locations/{location}/notebookExecutionJobs/{job_id}` + "notebookRuntimeTemplateResourceName": "A String", # The NotebookRuntimeTemplate to source compute configuration from. + "scheduleResourceName": "A String", # Output only. The Schedule resource name if this job is triggered by one. Format: `projects/{project_id}/locations/{location}/schedules/{schedule_id}` + "serviceAccount": "A String", # The service account to run the execution as. + "status": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # Output only. Populated when the NotebookExecutionJob is completed. When there is an error during notebook execution, the error details are populated. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "updateTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was most recently updated. +} + + notebookExecutionJobId: string, Optional. User specified ID for the NotebookExecutionJob. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+
delete(name, x__xgafv=None)
Deletes a NotebookExecutionJob. @@ -186,23 +265,6 @@Method Details
{ # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -267,23 +329,6 @@Method Details
"notebookExecutionJobs": [ # List of NotebookExecutionJobs in the requested page. { # NotebookExecutionJob represents an instance of a notebook execution. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimeTemplates.html b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimeTemplates.html index e4cb277a091..d59005b3fe6 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimeTemplates.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimeTemplates.html @@ -95,6 +95,9 @@Instance Methods
Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a NotebookRuntimeTemplate.
setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors.
@@ -124,6 +127,9 @@Method Details
}, "description": "A String", # The description of the NotebookRuntimeTemplate. "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. @@ -153,13 +159,6 @@Method Details
"A String", ], "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Optional. Reservation Affinity of the notebook runtime template. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. @@ -254,6 +253,9 @@Method Details
}, "description": "A String", # The description of the NotebookRuntimeTemplate. "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. @@ -283,13 +285,6 @@Method Details
"A String", ], "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Optional. Reservation Affinity of the notebook runtime template. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. @@ -363,6 +358,9 @@Method Details
}, "description": "A String", # The description of the NotebookRuntimeTemplate. "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. @@ -392,13 +390,6 @@Method Details
"A String", ], "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Optional. Reservation Affinity of the notebook runtime template. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. @@ -423,6 +414,119 @@Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates a NotebookRuntimeTemplate. + +Args: + name: string, The resource name of the NotebookRuntimeTemplate. (required) + body: object, The request body. + The object takes the form of: + +{ # A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template. + "createTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was created. + "dataPersistentDiskSpec": { # Represents the spec of persistent disk options. # Optional. The specification of persistent disk attached to the runtime as data disk storage. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + "description": "A String", # The description of the NotebookRuntimeTemplate. + "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, + "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. + "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. + "eucDisabled": True or False, # Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate. + }, + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, + "isDefault": True or False, # Output only. The default template to use if not specified. + "labels": { # The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. + "a_key": "A String", + }, + "machineSpec": { # Specification of a single machine. # Optional. Immutable. The specification of a single machine for the template. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "name": "A String", # The resource name of the NotebookRuntimeTemplate. + "networkSpec": { # Network spec. # Optional. Network spec. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "networkTags": [ # Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). + "A String", + ], + "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. + "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. + "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. + "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. + }, + "updateTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated. +} + + updateMask: string, Required. The update mask applies to the resource. For the `FieldMask` definition, see google.protobuf.FieldMask. Input format: `{paths: "${updated_filed}"}` Updatable fields: * `encryption_spec.kms_key_name` + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A template that specifies runtime configurations such as machine type, runtime version, network configurations, etc. Multiple runtimes can be created from a runtime template. + "createTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was created. + "dataPersistentDiskSpec": { # Represents the spec of persistent disk options. # Optional. The specification of persistent disk attached to the runtime as data disk storage. + "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). + "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) + }, + "description": "A String", # The description of the NotebookRuntimeTemplate. + "displayName": "A String", # Required. The display name of the NotebookRuntimeTemplate. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, + "etag": "A String", # Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "eucConfig": { # The euc configuration of NotebookRuntimeTemplate. # EUC configuration of the NotebookRuntimeTemplate. + "bypassActasCheck": True or False, # Output only. Whether ActAs check is bypassed for service account attached to the VM. If false, we need ActAs check for the default Compute Engine Service account. When a Runtime is created, a VM is allocated using Default Compute Engine Service Account. Any user requesting to use this Runtime requires Service Account User (ActAs) permission over this SA. If true, Runtime owner is using EUC and does not require the above permission as VM no longer use default Compute Engine SA, but a P4SA. + "eucDisabled": True or False, # Input only. Whether EUC is disabled in this NotebookRuntimeTemplate. In proto3, the default value of a boolean is false. In this way, by default EUC will be enabled for NotebookRuntimeTemplate. + }, + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # The idle shutdown configuration of NotebookRuntimeTemplate. This config will only be set when idle shutdown is enabled. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, + "isDefault": True or False, # Output only. The default template to use if not specified. + "labels": { # The labels with user-defined metadata to organize the NotebookRuntimeTemplates. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. See https://goo.gl/xmQnxf for more information and examples of labels. + "a_key": "A String", + }, + "machineSpec": { # Specification of a single machine. # Optional. Immutable. The specification of a single machine for the template. + "acceleratorCount": 42, # The number of accelerators to attach to the machine. + "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. + "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. + "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). + }, + "name": "A String", # The resource name of the NotebookRuntimeTemplate. + "networkSpec": { # Network spec. # Optional. Network spec. + "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. + "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) + "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` + }, + "networkTags": [ # Optional. The Compute Engine tags to add to runtime (see [Tagging instances](https://cloud.google.com/vpc/docs/add-remove-network-tags)). + "A String", + ], + "notebookRuntimeType": "A String", # Optional. Immutable. The type of the notebook runtime template. + "serviceAccount": "A String", # The service account that the runtime workload runs as. You can use any service account within the same project, but you must have the service account user permission to use the instance. If not specified, the [Compute Engine default service account](https://cloud.google.com/compute/docs/access/service-accounts#default_service_account) is used. + "shieldedVmConfig": { # A set of Shielded Instance options. See [Images using supported Shielded VM features](https://cloud.google.com/compute/docs/instances/modifying-shielded-vm). # Optional. Immutable. Runtime Shielded VM spec. + "enableSecureBoot": True or False, # Defines whether the instance has [Secure Boot](https://cloud.google.com/compute/shielded-vm/docs/shielded-vm#secure-boot) enabled. Secure Boot helps ensure that the system only runs authentic software by verifying the digital signature of all boot components, and halting the boot process if signature verification fails. + }, + "updateTime": "A String", # Output only. Timestamp when this NotebookRuntimeTemplate was most recently updated. +}+
setIamPolicy(resource, body=None, x__xgafv=None)
Sets the access control policy on the specified resource. Replaces any existing policy. Can return `NOT_FOUND`, `INVALID_ARGUMENT`, and `PERMISSION_DENIED` errors. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimes.html b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimes.html index c47321353de..ff8efaee4ca 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimes.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.notebookRuntimes.html @@ -119,8 +119,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -134,13 +141,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -269,8 +269,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -284,13 +291,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. @@ -327,8 +327,15 @@Method Details
"createTime": "A String", # Output only. Timestamp when this NotebookRuntime was created. "description": "A String", # The description of the NotebookRuntime. "displayName": "A String", # Required. The display name of the NotebookRuntime. The name can be up to 128 characters long and can consist of any UTF-8 characters. + "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Output only. Customer-managed encryption key spec for the notebook runtime. + "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. + }, "expirationTime": "A String", # Output only. Timestamp when this NotebookRuntime will be expired: 1. System Predefined NotebookRuntime: 24 hours after creation. After expiration, system predifined runtime will be deleted. 2. User created NotebookRuntime: 6 months after last upgrade. After expiration, user created runtime will be stopped and allowed for upgrade. "healthState": "A String", # Output only. The health state of the NotebookRuntime. + "idleShutdownConfig": { # The idle shutdown configuration of NotebookRuntimeTemplate, which contains the idle_timeout as required field. # Output only. The idle shutdown configuration of the notebook runtime. + "idleShutdownDisabled": True or False, # Whether Idle Shutdown is disabled in this NotebookRuntimeTemplate. + "idleTimeout": "A String", # Required. Duration is accurate to the second. In Notebook, Idle Timeout is accurate to minute so the range of idle_timeout (second) is: 10 * 60 ~ 1440 * 60. + }, "isUpgradable": True or False, # Output only. Whether NotebookRuntime is upgradable. "labels": { # The labels with user-defined metadata to organize your NotebookRuntime. Label keys and values can be no longer than 64 characters (Unicode codepoints), can only contain lowercase letters, numeric characters, underscores and dashes. International characters are allowed. No more than 64 user labels can be associated with one NotebookRuntime (System labels are excluded). See https://goo.gl/xmQnxf for more information and examples of labels. System reserved label keys are prefixed with "aiplatform.googleapis.com/" and are immutable. Following system labels exist for NotebookRuntime: * "aiplatform.googleapis.com/notebook_runtime_gce_instance_id": output only, its value is the Compute Engine instance id. * "aiplatform.googleapis.com/colab_enterprise_entry_service": its value is either "bigquery" or "vertex"; if absent, it should be "vertex". This is to describe the entry service, either BigQuery or Vertex. "a_key": "A String", @@ -342,13 +349,6 @@Method Details
}, "notebookRuntimeType": "A String", # Output only. The type of the notebook runtime. "proxyUri": "A String", # Output only. The proxy endpoint used to access the NotebookRuntime. - "reservationAffinity": { # Notebook Reservation Affinity for consuming Zonal reservation. # Output only. Reservation Affinity of the notebook runtime. - "consumeReservationType": "A String", # Required. Specifies the type of reservation from which this instance can consume resources: RESERVATION_ANY (default), RESERVATION_SPECIFIC, or RESERVATION_NONE. See Consuming reserved instances for examples. - "key": "A String", # Optional. Corresponds to the label key of a reservation resource. To target a RESERVATION_SPECIFIC by name, use compute.googleapis.com/reservation-name as the key and specify the name of your reservation as its value. - "values": [ # Optional. Corresponds to the label values of a reservation resource. This must be the full path name of Reservation. - "A String", - ], - }, "runtimeState": "A String", # Output only. The runtime (instance) state of the NotebookRuntime. "runtimeUser": "A String", # Required. The user email of the NotebookRuntime. "satisfiesPzi": True or False, # Output only. Reserved for future use. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html index 20cd2d3712e..cde13c36aee 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.publishers.models.html @@ -218,6 +218,7 @@Method Details
The object takes the form of: { # Request message for [PredictionService.GenerateContent]. + "cachedContent": "A String", # Optional. The name of the cached content used as context to serve the prediction. Note: only used in explicit caching, where users can have control over caching (e.g. what content to cache) and enjoy guaranteed cost savings. Format: `projects/{project}/locations/{location}/cachedContents/{cachedContent}` "contents": [ # Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request. { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. @@ -258,7 +259,34 @@Method Details
"maxOutputTokens": 42, # Optional. The maximum number of output tokens to generate per message. "presencePenalty": 3.14, # Optional. Positive penalties. "responseMimeType": "A String", # Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. - "responseStyle": "A String", # Optional. Control Three levels of creativity in the model output. Default: RESPONSE_STYLE_BALANCED + "responseSchema": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. The `Schema` object allows the definition of input and output data types. These types can be objects, but also primitives and arrays. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). If set, a compatible response_mime_type must also be set. Compatible mimetypes: `application/json`: Schema for JSON response. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, "stopSequences": [ # Optional. Stop sequences. "A String", ], @@ -377,6 +405,8 @@Method Details
}, }, ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. @@ -819,6 +849,7 @@Method Details
The object takes the form of: { # Request message for [PredictionService.GenerateContent]. + "cachedContent": "A String", # Optional. The name of the cached content used as context to serve the prediction. Note: only used in explicit caching, where users can have control over caching (e.g. what content to cache) and enjoy guaranteed cost savings. Format: `projects/{project}/locations/{location}/cachedContents/{cachedContent}` "contents": [ # Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries, this is a repeated field that contains conversation history + latest request. { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. @@ -859,7 +890,34 @@Method Details
"maxOutputTokens": 42, # Optional. The maximum number of output tokens to generate per message. "presencePenalty": 3.14, # Optional. Positive penalties. "responseMimeType": "A String", # Optional. Output response mimetype of the generated candidate text. Supported mimetype: - `text/plain`: (default) Text output. - `application/json`: JSON response in the candidates. The model needs to be prompted to output the appropriate response type, otherwise the behavior is undefined. This is a preview feature. - "responseStyle": "A String", # Optional. Control Three levels of creativity in the model output. Default: RESPONSE_STYLE_BALANCED + "responseSchema": { # Schema is used to define the format of input/output data. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). More fields may be added in the future as needed. # Optional. The `Schema` object allows the definition of input and output data types. These types can be objects, but also primitives and arrays. Represents a select subset of an [OpenAPI 3.0 schema object](https://spec.openapis.org/oas/v3.0.3#schema). If set, a compatible response_mime_type must also be set. Compatible mimetypes: `application/json`: Schema for JSON response. + "default": "", # Optional. Default value of the data. + "description": "A String", # Optional. The description of the data. + "enum": [ # Optional. Possible values of the element of Type.STRING with enum format. For example we can define an Enum Direction as : {type:STRING, format:enum, enum:["EAST", NORTH", "SOUTH", "WEST"]} + "A String", + ], + "example": "", # Optional. Example of the object. Will only populated when the object is the root. + "format": "A String", # Optional. The format of the data. Supported formats: for NUMBER type: "float", "double" for INTEGER type: "int32", "int64" for STRING type: "email", "byte", etc + "items": # Object with schema name: GoogleCloudAiplatformV1beta1Schema # Optional. SCHEMA FIELDS FOR TYPE ARRAY Schema of the elements of Type.ARRAY. + "maxItems": "A String", # Optional. Maximum number of the elements for Type.ARRAY. + "maxLength": "A String", # Optional. Maximum length of the Type.STRING + "maxProperties": "A String", # Optional. Maximum number of the properties for Type.OBJECT. + "maximum": 3.14, # Optional. Maximum value of the Type.INTEGER and Type.NUMBER + "minItems": "A String", # Optional. Minimum number of the elements for Type.ARRAY. + "minLength": "A String", # Optional. SCHEMA FIELDS FOR TYPE STRING Minimum length of the Type.STRING + "minProperties": "A String", # Optional. Minimum number of the properties for Type.OBJECT. + "minimum": 3.14, # Optional. SCHEMA FIELDS FOR TYPE INTEGER and NUMBER Minimum value of the Type.INTEGER and Type.NUMBER + "nullable": True or False, # Optional. Indicates if the value may be null. + "pattern": "A String", # Optional. Pattern of the Type.STRING to restrict a string to a regular expression. + "properties": { # Optional. SCHEMA FIELDS FOR TYPE OBJECT Properties of Type.OBJECT. + "a_key": # Object with schema name: GoogleCloudAiplatformV1beta1Schema + }, + "required": [ # Optional. Required properties of Type.OBJECT. + "A String", + ], + "title": "A String", # Optional. The title of the Schema. + "type": "A String", # Optional. The type of the data. + }, "stopSequences": [ # Optional. Stop sequences. "A String", ], @@ -978,6 +1036,8 @@Method Details
}, }, ], + "googleSearchRetrieval": { # Tool to retrieve public web data for grounding, powered by Google. # Optional. GoogleSearchRetrieval tool type. Specialized retrieval tool that is powered by Google search. + }, "retrieval": { # Defines a retrieval tool that model can call to access external knowledge. # Optional. Retrieval tool type. System will always execute the provided retrieval tool(s) to get external knowledge to answer the prompt. Retrieval results are presented to the model for generation. "disableAttribution": True or False, # Optional. Disable using the result from this tool in detecting grounding attribution. This does not affect how the result is given to the model for generation. "vertexAiSearch": { # Retrieve from Vertex AI Search datastore for grounding. See https://cloud.google.com/vertex-ai-search-and-conversation # Set to use data source powered by Vertex AI Search. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html index b7501e96508..fa8937a2862 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.reasoningEngines.html @@ -97,6 +97,9 @@Instance Methods
Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a reasoning engine.
query(name, body=None, x__xgafv=None)
Queries using a reasoning engine.
@@ -297,6 +300,67 @@Method Details
patch(name, body=None, updateMask=None, x__xgafv=None)
+ Updates a reasoning engine. + +Args: + name: string, Identifier. The resource name of the ReasoningEngine. (required) + body: object, The request body. + The object takes the form of: + +{ # ReasoningEngine provides a customizable runtime for models to determine which actions to take and in which order. + "createTime": "A String", # Output only. Timestamp when this ReasoningEngine was created. + "description": "A String", # Optional. The description of the ReasoningEngine. + "displayName": "A String", # Required. The display name of the ReasoningEngine. + "etag": "A String", # Optional. Used to perform consistent read-modify-write updates. If not set, a blind "overwrite" update happens. + "name": "A String", # Identifier. The resource name of the ReasoningEngine. + "spec": { # ReasoningEngine configurations # Required. Configurations of the ReasoningEngine + "classMethods": [ # Optional. Declarations for object class methods. + { + "a_key": "", # Properties of the object. + }, + ], + "packageSpec": { # User provided package spec like pickled object and package requirements. # Required. User provided package spec of the ReasoningEngine. + "dependencyFilesGcsUri": "A String", # Optional. The Cloud Storage URI of the dependency files in tar.gz format. + "pickleObjectGcsUri": "A String", # Optional. The Cloud Storage URI of the pickled python object. + "pythonVersion": "A String", # Optional. The Python version. Currently support 3.8, 3.9, 3.10, 3.11. If not specified, default value is 3.10. + "requirementsGcsUri": "A String", # Optional. The Cloud Storage URI of the `requirements.txt` file + }, + }, + "updateTime": "A String", # Output only. Timestamp when this ReasoningEngine was most recently updated. +} + + updateMask: string, Required. Mask specifying which fields to update. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+
query(name, body=None, x__xgafv=None)
Queries using a reasoning engine. diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html index c55afbb193f..3b47bbc3a91 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.schedules.html @@ -436,23 +436,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1052,23 +1035,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -1710,23 +1676,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -2340,23 +2289,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -2974,23 +2906,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` @@ -3591,23 +3506,6 @@Method Details
"createNotebookExecutionJobRequest": { # Request message for [NotebookService.CreateNotebookExecutionJob] # Request for NotebookService.CreateNotebookExecutionJob. "notebookExecutionJob": { # NotebookExecutionJob represents an instance of a notebook execution. # Required. The NotebookExecutionJob to create. "createTime": "A String", # Output only. Timestamp when this NotebookExecutionJob was created. - "customEnvironmentSpec": { # Compute configuration to use for an execution job. # The custom compute configuration for an execution job. - "machineSpec": { # Specification of a single machine. # The specification of a single machine for the execution job. - "acceleratorCount": 42, # The number of accelerators to attach to the machine. - "acceleratorType": "A String", # Immutable. The type of accelerator(s) that may be attached to the machine as per accelerator_count. - "machineType": "A String", # Immutable. The type of the machine. See the [list of machine types supported for prediction](https://cloud.google.com/vertex-ai/docs/predictions/configure-compute#machine-types) See the [list of machine types supported for custom training](https://cloud.google.com/vertex-ai/docs/training/configure-compute#machine-types). For DeployedModel this field is optional, and the default value is `n1-standard-2`. For BatchPredictionJob or as part of WorkerPoolSpec this field is required. - "tpuTopology": "A String", # Immutable. The topology of the TPUs. Corresponds to the TPU topologies available from GKE. (Example: tpu_topology: "2x2x1"). - }, - "networkSpec": { # Network spec. # The network configuration to use for the execution job. - "enableInternetAccess": True or False, # Whether to enable public internet access. Default false. - "network": "A String", # The full name of the Google Compute Engine [network](https://cloud.google.com//compute/docs/networks-and-firewalls#networks) - "subnetwork": "A String", # The name of the subnet that this instance is in. Format: `projects/{project_id_or_number}/regions/{region}/subnetworks/{subnetwork_id}` - }, - "persistentDiskSpec": { # Represents the spec of persistent disk options. # The specification of a persistent disk to attach for the execution job. - "diskSizeGb": "A String", # Size in GB of the disk (default is 100GB). - "diskType": "A String", # Type of the disk (default is "pd-standard"). Valid values: "pd-ssd" (Persistent Disk Solid State Drive) "pd-standard" (Persistent Disk Hard Disk Drive) "pd-balanced" (Balanced Persistent Disk) "pd-extreme" (Extreme Persistent Disk) - }, - }, "dataformRepositorySource": { # The Dataform Repository containing the input notebook. # The Dataform Repository pointing to a single file notebook repository. "commitSha": "A String", # The commit SHA to read repository with. If unset, the file will be read at HEAD. "dataformRepositoryResourceName": "A String", # The resource name of the Dataform Repository. Format: `projects/{project_id}/locations/{location}/repositories/{repository_id}` diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html b/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html index 57650953c6f..d3f4b6c0810 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.trainingPipelines.html @@ -227,7 +227,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -245,7 +245,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -532,7 +532,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -550,7 +550,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -879,7 +879,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -897,7 +897,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1198,7 +1198,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -1216,7 +1216,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], diff --git a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html index 6dea72a0c3c..132e8d6bbaf 100644 --- a/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html +++ b/docs/dyn/aiplatform_v1beta1.projects.locations.tuningJobs.html @@ -135,6 +135,19 @@Method Details
"baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. + "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. + "baseTeacherModel": "A String", # The base teacher model that is being distilled, e.g., "gemini-1.0-pro-002". + "hyperParameters": { # Hyperparameters for Distillation. # Optional. Hyperparameters for Distillation. + "adapterSize": "A String", # Optional. Adapter size for distillation. + "epochCount": "A String", # Optional. Number of complete passes the model makes over the entire training dataset during training. + "learningRateMultiplier": 3.14, # Optional. Multiplier for adjusting the default learning rate. + }, + "pipelineRootDirectory": "A String", # Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the distillation pipeline. It is used by the system to generate the paths of output artifacts. + "studentModel": "A String", # The student model that is being tuned, e.g., "google/gemma-2b-it". + "trainingDatasetUri": "A String", # Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. + "tunedTeacherModelSource": "A String", # The resource name of the Tuned teacher model. Format: `projects/{project}/locations/{location}/models/{model}`. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. + }, "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. }, @@ -153,6 +166,23 @@Method Details
"a_key": "A String", }, "name": "A String", # Output only. Identifier. Resource name of a TuningJob. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}` + "pipelineJob": "A String", # Output only. The resource name of the PipelineJob associated with the TuningJob. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`. + "reinforcementLearningSpec": { # Tuning Spec for Reinforcement Learning. # Tuning Spec for Reinforcement Learning. + "hyperParameters": { # Hyperparameters for Reinforcement Learning. # Optional. Additional hyper-parameters to use during tuning. + "epochCount": "A String", # Optional. Number of training epoches for the tuning job. + "humanFeedbackConfig": { # Configures Reinforcement Learning to use human feedback during tuning. # Configures Reinforcement Learning to use human feedback for preference data during tuning. + "preferenceDatasetUri": "A String", # Required. Cloud Storage path to human preference data. + }, + "klCoefficient": 3.14, # Optional. KL divergence coefficient for Reinforcement Learning. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for Reinforcement Learning. + "rewardModelTrainingConfig": { # Configures Reinforcement Learning to learn preference by training a reward model. # Configures Reinforcement Learning to train a reward model to learn preference. + "epochCount": "A String", # Optional. Number of training epoches for the reward model training job. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for reward model training. + }, + }, + "promptDatasetUri": "A String", # Required. Cloud Storage path to the prompt dataset to use during Reinforcement Learning. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to the validation dataset to use during Reinforcement Learning. + }, "startTime": "A String", # Output only. Time when the TuningJob for the first time entered the `JOB_STATE_RUNNING` state. "state": "A String", # Output only. The detailed state of the job. "supervisedTuningSpec": { # Tuning Spec for Supervised Tuning. # Tuning Spec for Supervised Fine Tuning. @@ -170,6 +200,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -274,6 +572,19 @@Method Details
"baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. + "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. + "baseTeacherModel": "A String", # The base teacher model that is being distilled, e.g., "gemini-1.0-pro-002". + "hyperParameters": { # Hyperparameters for Distillation. # Optional. Hyperparameters for Distillation. + "adapterSize": "A String", # Optional. Adapter size for distillation. + "epochCount": "A String", # Optional. Number of complete passes the model makes over the entire training dataset during training. + "learningRateMultiplier": 3.14, # Optional. Multiplier for adjusting the default learning rate. + }, + "pipelineRootDirectory": "A String", # Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the distillation pipeline. It is used by the system to generate the paths of output artifacts. + "studentModel": "A String", # The student model that is being tuned, e.g., "google/gemma-2b-it". + "trainingDatasetUri": "A String", # Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. + "tunedTeacherModelSource": "A String", # The resource name of the Tuned teacher model. Format: `projects/{project}/locations/{location}/models/{model}`. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. + }, "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. }, @@ -292,6 +603,23 @@Method Details
"a_key": "A String", }, "name": "A String", # Output only. Identifier. Resource name of a TuningJob. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}` + "pipelineJob": "A String", # Output only. The resource name of the PipelineJob associated with the TuningJob. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`. + "reinforcementLearningSpec": { # Tuning Spec for Reinforcement Learning. # Tuning Spec for Reinforcement Learning. + "hyperParameters": { # Hyperparameters for Reinforcement Learning. # Optional. Additional hyper-parameters to use during tuning. + "epochCount": "A String", # Optional. Number of training epoches for the tuning job. + "humanFeedbackConfig": { # Configures Reinforcement Learning to use human feedback during tuning. # Configures Reinforcement Learning to use human feedback for preference data during tuning. + "preferenceDatasetUri": "A String", # Required. Cloud Storage path to human preference data. + }, + "klCoefficient": 3.14, # Optional. KL divergence coefficient for Reinforcement Learning. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for Reinforcement Learning. + "rewardModelTrainingConfig": { # Configures Reinforcement Learning to learn preference by training a reward model. # Configures Reinforcement Learning to train a reward model to learn preference. + "epochCount": "A String", # Optional. Number of training epoches for the reward model training job. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for reward model training. + }, + }, + "promptDatasetUri": "A String", # Required. Cloud Storage path to the prompt dataset to use during Reinforcement Learning. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to the validation dataset to use during Reinforcement Learning. + }, "startTime": "A String", # Output only. Time when the TuningJob for the first time entered the `JOB_STATE_RUNNING` state. "state": "A String", # Output only. The detailed state of the job. "supervisedTuningSpec": { # Tuning Spec for Supervised Tuning. # Tuning Spec for Supervised Fine Tuning. @@ -309,6 +637,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -420,6 +1016,19 @@Method Details
"baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. + "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. + "baseTeacherModel": "A String", # The base teacher model that is being distilled, e.g., "gemini-1.0-pro-002". + "hyperParameters": { # Hyperparameters for Distillation. # Optional. Hyperparameters for Distillation. + "adapterSize": "A String", # Optional. Adapter size for distillation. + "epochCount": "A String", # Optional. Number of complete passes the model makes over the entire training dataset during training. + "learningRateMultiplier": 3.14, # Optional. Multiplier for adjusting the default learning rate. + }, + "pipelineRootDirectory": "A String", # Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the distillation pipeline. It is used by the system to generate the paths of output artifacts. + "studentModel": "A String", # The student model that is being tuned, e.g., "google/gemma-2b-it". + "trainingDatasetUri": "A String", # Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. + "tunedTeacherModelSource": "A String", # The resource name of the Tuned teacher model. Format: `projects/{project}/locations/{location}/models/{model}`. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. + }, "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. }, @@ -438,6 +1047,23 @@Method Details
"a_key": "A String", }, "name": "A String", # Output only. Identifier. Resource name of a TuningJob. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}` + "pipelineJob": "A String", # Output only. The resource name of the PipelineJob associated with the TuningJob. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`. + "reinforcementLearningSpec": { # Tuning Spec for Reinforcement Learning. # Tuning Spec for Reinforcement Learning. + "hyperParameters": { # Hyperparameters for Reinforcement Learning. # Optional. Additional hyper-parameters to use during tuning. + "epochCount": "A String", # Optional. Number of training epoches for the tuning job. + "humanFeedbackConfig": { # Configures Reinforcement Learning to use human feedback during tuning. # Configures Reinforcement Learning to use human feedback for preference data during tuning. + "preferenceDatasetUri": "A String", # Required. Cloud Storage path to human preference data. + }, + "klCoefficient": 3.14, # Optional. KL divergence coefficient for Reinforcement Learning. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for Reinforcement Learning. + "rewardModelTrainingConfig": { # Configures Reinforcement Learning to learn preference by training a reward model. # Configures Reinforcement Learning to train a reward model to learn preference. + "epochCount": "A String", # Optional. Number of training epoches for the reward model training job. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for reward model training. + }, + }, + "promptDatasetUri": "A String", # Required. Cloud Storage path to the prompt dataset to use during Reinforcement Learning. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to the validation dataset to use during Reinforcement Learning. + }, "startTime": "A String", # Output only. Time when the TuningJob for the first time entered the `JOB_STATE_RUNNING` state. "state": "A String", # Output only. The detailed state of the job. "supervisedTuningSpec": { # Tuning Spec for Supervised Tuning. # Tuning Spec for Supervised Fine Tuning. @@ -455,6 +1081,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. @@ -572,6 +1466,19 @@Method Details
"baseModel": "A String", # The base model that is being tuned, e.g., "gemini-1.0-pro-002". "createTime": "A String", # Output only. Time when the TuningJob was created. "description": "A String", # Optional. The description of the TuningJob. + "distillationSpec": { # Tuning Spec for Distillation. # Tuning Spec for Distillation. + "baseTeacherModel": "A String", # The base teacher model that is being distilled, e.g., "gemini-1.0-pro-002". + "hyperParameters": { # Hyperparameters for Distillation. # Optional. Hyperparameters for Distillation. + "adapterSize": "A String", # Optional. Adapter size for distillation. + "epochCount": "A String", # Optional. Number of complete passes the model makes over the entire training dataset during training. + "learningRateMultiplier": 3.14, # Optional. Multiplier for adjusting the default learning rate. + }, + "pipelineRootDirectory": "A String", # Required. A path in a Cloud Storage bucket, which will be treated as the root output directory of the distillation pipeline. It is used by the system to generate the paths of output artifacts. + "studentModel": "A String", # The student model that is being tuned, e.g., "google/gemma-2b-it". + "trainingDatasetUri": "A String", # Required. Cloud Storage path to file containing training dataset for tuning. The dataset must be formatted as a JSONL file. + "tunedTeacherModelSource": "A String", # The resource name of the Tuned teacher model. Format: `projects/{project}/locations/{location}/models/{model}`. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to file containing validation dataset for tuning. The dataset must be formatted as a JSONL file. + }, "encryptionSpec": { # Represents a customer-managed encryption key spec that can be applied to a top-level resource. # Customer-managed encryption key options for a TuningJob. If this is set, then all resources created by the TuningJob will be encrypted with the provided encryption key. "kmsKeyName": "A String", # Required. The Cloud KMS resource identifier of the customer managed encryption key used to protect a resource. Has the form: `projects/my-project/locations/my-region/keyRings/my-kr/cryptoKeys/my-key`. The key needs to be in the same region as where the compute resource is created. }, @@ -590,6 +1497,23 @@Method Details
"a_key": "A String", }, "name": "A String", # Output only. Identifier. Resource name of a TuningJob. Format: `projects/{project}/locations/{location}/tuningJobs/{tuning_job}` + "pipelineJob": "A String", # Output only. The resource name of the PipelineJob associated with the TuningJob. Format: `projects/{project}/locations/{location}/pipelineJobs/{pipeline_job}`. + "reinforcementLearningSpec": { # Tuning Spec for Reinforcement Learning. # Tuning Spec for Reinforcement Learning. + "hyperParameters": { # Hyperparameters for Reinforcement Learning. # Optional. Additional hyper-parameters to use during tuning. + "epochCount": "A String", # Optional. Number of training epoches for the tuning job. + "humanFeedbackConfig": { # Configures Reinforcement Learning to use human feedback during tuning. # Configures Reinforcement Learning to use human feedback for preference data during tuning. + "preferenceDatasetUri": "A String", # Required. Cloud Storage path to human preference data. + }, + "klCoefficient": 3.14, # Optional. KL divergence coefficient for Reinforcement Learning. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for Reinforcement Learning. + "rewardModelTrainingConfig": { # Configures Reinforcement Learning to learn preference by training a reward model. # Configures Reinforcement Learning to train a reward model to learn preference. + "epochCount": "A String", # Optional. Number of training epoches for the reward model training job. + "learningRateMultiplier": 3.14, # Optional. Learning rate multiplier for reward model training. + }, + }, + "promptDatasetUri": "A String", # Required. Cloud Storage path to the prompt dataset to use during Reinforcement Learning. + "validationDatasetUri": "A String", # Optional. Cloud Storage path to the validation dataset to use during Reinforcement Learning. + }, "startTime": "A String", # Output only. Time when the TuningJob for the first time entered the `JOB_STATE_RUNNING` state. "state": "A String", # Output only. The detailed state of the job. "supervisedTuningSpec": { # Tuning Spec for Supervised Tuning. # Tuning Spec for Supervised Fine Tuning. @@ -607,6 +1531,274 @@Method Details
}, "tunedModelDisplayName": "A String", # Optional. The display name of the TunedModel. The name can be up to 128 characters long and can consist of any UTF-8 characters. "tuningDataStats": { # The tuning data statistic values for TuningJob. # Output only. The tuning data statistics associated with this TuningJob. + "distillationDataStats": { # Statistics computed for datasets used for distillation. # Statistics for distillation. + "trainingDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the training dataset. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, + "reinforcementLearningDataStats": { # Statistics computed for datasets used for reinforcement learning. # Statistics for reinforcement learning. + "preferenceDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the preference dataset. This can be either a human preference dataset or a preference dataset generated from AI feedback. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + "promptDatasetStats": { # Statistics computed over a tuning dataset. # Output only. Statistics computed for the prompt dataset used during reinforcement learning. + "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. + "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. + "tuningDatasetExampleCount": "A String", # Output only. Number of examples in the tuning dataset. + "tuningStepCount": "A String", # Output only. Number of tuning steps for this Tuning Job. + "userDatasetExamples": [ # Output only. Sample user messages in the training dataset uri. + { # The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn. + "parts": [ # Required. Ordered `Parts` that constitute a single message. Parts may have different IANA MIME types. + { # A datatype containing media that is part of a multi-part `Content` message. A `Part` consists of data which has an associated datatype. A `Part` can only contain one of the accepted types in `Part.data`. A `Part` must have a fixed IANA MIME type identifying the type and subtype of the media if `inline_data` or `file_data` field is filled with raw bytes. + "fileData": { # URI based data. # Optional. URI based data. + "fileUri": "A String", # Required. URI. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "functionCall": { # A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing the parameters and their values. # Optional. A predicted [FunctionCall] returned from the model that contains a string representing the [FunctionDeclaration.name] with the parameters and their values. + "args": { # Optional. Required. The function parameters and values in JSON object format. See [FunctionDeclaration.parameters] for parameter details. + "a_key": "", # Properties of the object. + }, + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name]. + }, + "functionResponse": { # The result output from a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function is used as context to the model. This should contain the result of a [FunctionCall] made based on model prediction. # Optional. The result output of a [FunctionCall] that contains a string representing the [FunctionDeclaration.name] and a structured JSON object containing any output from the function call. It is used as context to the model. + "name": "A String", # Required. The name of the function to call. Matches [FunctionDeclaration.name] and [FunctionCall.name]. + "response": { # Required. The function response in JSON object format. + "a_key": "", # Properties of the object. + }, + }, + "inlineData": { # Content blob. It's preferred to send as text directly rather than raw bytes. # Optional. Inlined bytes data. + "data": "A String", # Required. Raw bytes. + "mimeType": "A String", # Required. The IANA standard MIME type of the source data. + }, + "text": "A String", # Optional. Text part (can be code). + "videoMetadata": { # Metadata describes the input video content. # Optional. Video metadata. The metadata should only be specified while the video data is presented in inline_data or file_data. + "endOffset": "A String", # Optional. The end offset of the video. + "startOffset": "A String", # Optional. The start offset of the video. + }, + }, + ], + "role": "A String", # Optional. The producer of the content. Must be either 'user' or 'model'. Useful to set for multi-turn conversations, otherwise can be left blank or unset. + }, + ], + "userInputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user input tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userMessagePerExampleDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the messages per example. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + "userOutputTokenDistribution": { # Distribution computed over a tuning dataset. # Output only. Dataset distributions for the user output tokens. + "buckets": [ # Output only. Defines the histogram bucket. + { # Dataset bucket used to create a histogram for the distribution given a population of values. + "count": "A String", # Output only. Number of values in the bucket. + "left": 3.14, # Output only. Left bound of the bucket. + "right": 3.14, # Output only. Right bound of the bucket. + }, + ], + "max": 3.14, # Output only. The maximum of the population values. + "mean": 3.14, # Output only. The arithmetic mean of the values in the population. + "median": 3.14, # Output only. The median of the values in the population. + "min": 3.14, # Output only. The minimum of the population values. + "p5": 3.14, # Output only. The 5th percentile of the values in the population. + "p95": 3.14, # Output only. The 95th percentile of the values in the population. + "sum": 3.14, # Output only. Sum of a given population of values. + }, + }, + }, "supervisedTuningDataStats": { # Tuning data statistics for Supervised Tuning. # The SFT Tuning data stats. "totalBillableCharacterCount": "A String", # Output only. Number of billable characters in the tuning dataset. "totalTuningCharacterCount": "A String", # Output only. Number of tuning characters in the tuning dataset. diff --git a/docs/dyn/aiplatform_v1beta1.publishers.models.html b/docs/dyn/aiplatform_v1beta1.publishers.models.html index 9e78faa3509..9b1091ee5e2 100644 --- a/docs/dyn/aiplatform_v1beta1.publishers.models.html +++ b/docs/dyn/aiplatform_v1beta1.publishers.models.html @@ -176,7 +176,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -194,7 +194,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -487,7 +487,7 @@Method Details
}, ], "healthProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes readiness probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], @@ -505,7 +505,7 @@Method Details
"predictRoute": "A String", # Immutable. HTTP path on the container to send prediction requests to. Vertex AI forwards requests sent using projects.locations.endpoints.predict to this path on the container's IP address and port. Vertex AI then returns the container's response in the API response. For example, if you set this field to `/foo`, then when Vertex AI receives a prediction request, it forwards the request body in a POST request to the `/foo` path on the port of your container specified by the first value of this `ModelContainerSpec`'s ports field. If you don't specify this field, it defaults to the following value when you deploy this Model to an Endpoint: /v1/endpoints/ENDPOINT/deployedModels/DEPLOYED_MODEL:predict The placeholders in this value are replaced as follows: * ENDPOINT: The last segment (following `endpoints/`)of the Endpoint.name][] field of the Endpoint where this Model has been deployed. (Vertex AI makes this value available to your container code as the [`AIP_ENDPOINT_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) * DEPLOYED_MODEL: DeployedModel.id of the `DeployedModel`. (Vertex AI makes this value available to your container code as the [`AIP_DEPLOYED_MODEL_ID` environment variable](https://cloud.google.com/vertex-ai/docs/predictions/custom-container-requirements#aip-variables).) "sharedMemorySizeMb": "A String", # Immutable. The amount of the VM memory to reserve as the shared memory for the model in megabytes. "startupProbe": { # Probe describes a health check to be performed against a container to determine whether it is alive or ready to receive traffic. # Immutable. Specification for Kubernetes startup probe. - "exec": { # ExecAction specifies a command to execute. # Exec specifies the action to take. + "exec": { # ExecAction specifies a command to execute. # ExecAction probes the health of a container by executing a command. "command": [ # Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy. "A String", ], diff --git a/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.eventEditRules.html b/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.eventEditRules.html new file mode 100644 index 00000000000..49160ce176f --- /dev/null +++ b/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.eventEditRules.html @@ -0,0 +1,116 @@ + + + +Google Analytics Admin API . properties . dataStreams . eventEditRules
+Instance Methods
++
+close()
Close httplib2 connections.
++
+reorder(parent, body=None, x__xgafv=None)
Changes the processing order of event edit rules on the specified stream.
+Method Details
+++ +close()
+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.html b/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.html index 3beb3dcec17..23e2ba6562b 100644 --- a/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.html +++ b/docs/dyn/analyticsadmin_v1alpha.properties.dataStreams.html @@ -79,6 +79,11 @@reorder(parent, body=None, x__xgafv=None)
+Changes the processing order of event edit rules on the specified stream. + +Args: + parent: string, Required. Example format: properties/123/dataStreams/456 (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for ReorderEventEditRules RPC. + "eventEditRules": [ # Required. EventEditRule resource names for the specified data stream, in the needed processing order. All EventEditRules for the stream must be present in the list. + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } +}+Instance Methods
Returns the eventCreateRules Resource.
++
+eventEditRules()
+Returns the eventEditRules Resource.
+ diff --git a/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.eventEditRules.html b/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.eventEditRules.html new file mode 100644 index 00000000000..5ecf0b96e50 --- /dev/null +++ b/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.eventEditRules.html @@ -0,0 +1,116 @@ + + + +Google Analytics Admin API . properties . dataStreams . eventEditRules
+Instance Methods
++
+close()
Close httplib2 connections.
++
+reorder(parent, body=None, x__xgafv=None)
Changes the processing order of event edit rules on the specified stream.
+Method Details
+++ +close()
+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.html b/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.html index ffa13534241..b3a1636bb64 100644 --- a/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.html +++ b/docs/dyn/analyticsadmin_v1beta.properties.dataStreams.html @@ -74,6 +74,11 @@reorder(parent, body=None, x__xgafv=None)
+Changes the processing order of event edit rules on the specified stream. + +Args: + parent: string, Required. Example format: properties/123/dataStreams/456 (required) + body: object, The request body. + The object takes the form of: + +{ # Request message for ReorderEventEditRules RPC. + "eventEditRules": [ # Required. EventEditRule resource names for the specified data stream, in the needed processing order. All EventEditRules for the stream must be present in the list. + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } +}+Google Analytics Admin API . properties . dataStreams
Instance Methods
++
+eventEditRules()
+Returns the eventEditRules Resource.
+ diff --git a/docs/dyn/androidmanagement_v1.enterprises.html b/docs/dyn/androidmanagement_v1.enterprises.html index 6a3ab870e05..6f3a39da71d 100644 --- a/docs/dyn/androidmanagement_v1.enterprises.html +++ b/docs/dyn/androidmanagement_v1.enterprises.html @@ -159,6 +159,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. @@ -220,6 +223,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. @@ -302,6 +308,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. @@ -374,6 +383,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. @@ -450,6 +462,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. @@ -508,6 +523,9 @@Method Details
"A String", ], "enterpriseDisplayName": "A String", # The name of the enterprise displayed to users. This field has a maximum length of 100 characters. + "googleAuthenticationSettings": { # Contains settings for Google-provided user authentication. # Settings for Google-provided user authentication. + "googleAuthenticationRequired": "A String", # Output only. Whether users need to be authenticated by Google during the enrollment process. IT admin can specify if Google authentication is enabled for the enterprise for knowledge worker devices. This value can be set only via the Google Admin Console. Google authentication can be used with signin_url In the case where Google authentication is required and a signin_url is specified, Google authentication will be launched before signin_url. + }, "logo": { # Data hosted at an external location. The data is to be downloaded by Android Device Policy and verified against the hash. # An image displayed as a logo during device provisioning. Supported types are: image/bmp, image/gif, image/x-ico, image/jpeg, image/png, image/webp, image/vnd.wap.wbmp, image/x-adobe-dng. "sha256Hash": "A String", # The base-64 encoded SHA-256 hash of the content hosted at url. If the content doesn't match this hash, Android Device Policy won't use the data. "url": "A String", # The absolute URL to the data, which must use either the http or https scheme. Android Device Policy doesn't provide any credentials in the GET request, so the URL must be publicly accessible. Including a long, random component in the URL may be used to prevent attackers from discovering the URL. diff --git a/docs/dyn/androidmanagement_v1.enterprises.policies.html b/docs/dyn/androidmanagement_v1.enterprises.policies.html index d28a511df3f..b98b98a800c 100644 --- a/docs/dyn/androidmanagement_v1.enterprises.policies.html +++ b/docs/dyn/androidmanagement_v1.enterprises.policies.html @@ -201,6 +201,7 @@Method Details
"policy": "A String", # The policy for granting the permission. }, ], + "userControlSettings": "A String", # Optional. Specifies whether user control is permitted for the app. User control includes user actions like force-stopping and clearing app data. Supported on Android 11 and above. "workProfileWidgets": "A String", # Specifies whether the app installed in the work profile is allowed to add widgets to the home screen. }, ], @@ -620,6 +621,7 @@Method Details
"policy": "A String", # The policy for granting the permission. }, ], + "userControlSettings": "A String", # Optional. Specifies whether user control is permitted for the app. User control includes user actions like force-stopping and clearing app data. Supported on Android 11 and above. "workProfileWidgets": "A String", # Specifies whether the app installed in the work profile is allowed to add widgets to the home screen. }, ], @@ -1045,6 +1047,7 @@Method Details
"policy": "A String", # The policy for granting the permission. }, ], + "userControlSettings": "A String", # Optional. Specifies whether user control is permitted for the app. User control includes user actions like force-stopping and clearing app data. Supported on Android 11 and above. "workProfileWidgets": "A String", # Specifies whether the app installed in the work profile is allowed to add widgets to the home screen. }, ], @@ -1453,6 +1456,7 @@Method Details
"policy": "A String", # The policy for granting the permission. }, ], + "userControlSettings": "A String", # Optional. Specifies whether user control is permitted for the app. User control includes user actions like force-stopping and clearing app data. Supported on Android 11 and above. "workProfileWidgets": "A String", # Specifies whether the app installed in the work profile is allowed to add widgets to the home screen. }, ], diff --git a/docs/dyn/backupdr_v1.projects.locations.managementServers.html b/docs/dyn/backupdr_v1.projects.locations.managementServers.html index 1f2d71a56cd..acdb5d0c7d9 100644 --- a/docs/dyn/backupdr_v1.projects.locations.managementServers.html +++ b/docs/dyn/backupdr_v1.projects.locations.managementServers.html @@ -138,6 +138,8 @@Method Details
}, ], "oauth2ClientId": "A String", # Output only. The OAuth 2.0 client id is required to make API calls to the BackupDR instance API of this ManagementServer. This is the value that should be provided in the ‘aud’ field of the OIDC ID Token (see openid specification https://openid.net/specs/openid-connect-core-1_0.html#IDToken). + "satisfiesPzi": True or False, # Output only. Reserved for future use. + "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The ManagementServer state. "type": "A String", # Optional. The type of the ManagementServer resource. "updateTime": "A String", # Output only. The time when the instance was updated. @@ -254,6 +256,8 @@Method Details
}, ], "oauth2ClientId": "A String", # Output only. The OAuth 2.0 client id is required to make API calls to the BackupDR instance API of this ManagementServer. This is the value that should be provided in the ‘aud’ field of the OIDC ID Token (see openid specification https://openid.net/specs/openid-connect-core-1_0.html#IDToken). + "satisfiesPzi": True or False, # Output only. Reserved for future use. + "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The ManagementServer state. "type": "A String", # Optional. The type of the ManagementServer resource. "updateTime": "A String", # Output only. The time when the instance was updated. @@ -358,6 +362,8 @@Method Details
}, ], "oauth2ClientId": "A String", # Output only. The OAuth 2.0 client id is required to make API calls to the BackupDR instance API of this ManagementServer. This is the value that should be provided in the ‘aud’ field of the OIDC ID Token (see openid specification https://openid.net/specs/openid-connect-core-1_0.html#IDToken). + "satisfiesPzi": True or False, # Output only. Reserved for future use. + "satisfiesPzs": True or False, # Output only. Reserved for future use. "state": "A String", # Output only. The ManagementServer state. "type": "A String", # Optional. The type of the ManagementServer resource. "updateTime": "A String", # Output only. The time when the instance was updated. diff --git a/docs/dyn/binaryauthorization_v1.projects.attestors.html b/docs/dyn/binaryauthorization_v1.projects.attestors.html index cae1ab19786..76fd2752913 100644 --- a/docs/dyn/binaryauthorization_v1.projects.attestors.html +++ b/docs/dyn/binaryauthorization_v1.projects.attestors.html @@ -129,7 +129,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. @@ -161,7 +161,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. @@ -217,7 +217,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. @@ -294,7 +294,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. @@ -432,7 +432,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. @@ -463,7 +463,7 @@Method Details
"updateTime": "A String", # Output only. Time when the attestor was last updated. "userOwnedGrafeasNote": { # An user owned Grafeas note references a Grafeas Attestation.Authority Note created by the user. # This specifies how an attestation will be read, and how it will be used during policy enforcement. "delegationServiceAccountEmail": "A String", # Output only. This field will contain the service account email address that this attestor will use as the principal when querying Container Analysis. Attestor administrators must grant this service account the IAM role needed to read attestations from the note_reference in Container Analysis (`containeranalysis.notes.occurrences.viewer`). This email address is fixed for the lifetime of the attestor, but callers should not make any other assumptions about the service account email; future versions may use an email based on a different naming pattern. - "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/*/notes/*`. This field may not be updated. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. + "noteReference": "A String", # Required. The Grafeas resource name of a Attestation.Authority Note, created by the user, in the format: `projects/[PROJECT_ID]/notes/*`. This field may not be updated. A project ID must be used, not a project number. An attestation by this attestor is stored as a Grafeas Attestation.Authority Occurrence that names a container image and that links to this Note. Grafeas is an external dependency. "publicKeys": [ # Optional. Public keys that verify attestations signed by this attestor. This field may be updated. If this field is non-empty, one of the specified public keys must verify that an attestation was signed by this attestor for the image specified in the admission request. If this field is empty, this attestor always returns that no valid attestations exist. { # An attestor public key that will be used to verify attestations signed by this attestor. "asciiArmoredPgpPublicKey": "A String", # ASCII-armored representation of a PGP public key, as the entire output by the command `gpg --export --armor foo@example.com` (either LF or CRLF line endings). When using this field, `id` should be left blank. The Binary Authorization API handlers will calculate the ID and fill it in automatically. Binary Authorization computes this ID as the OpenPGP RFC4880 V4 fingerprint, represented as upper-case hex. If `id` is provided by the caller, it will be overwritten by the API-calculated ID. diff --git a/docs/dyn/binaryauthorization_v1.projects.platforms.policies.html b/docs/dyn/binaryauthorization_v1.projects.platforms.policies.html index 876cde4d65e..e526d72aafb 100644 --- a/docs/dyn/binaryauthorization_v1.projects.platforms.policies.html +++ b/docs/dyn/binaryauthorization_v1.projects.platforms.policies.html @@ -156,7 +156,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, @@ -273,7 +273,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, @@ -414,7 +414,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, @@ -542,7 +542,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, @@ -676,7 +676,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, @@ -792,7 +792,7 @@Method Details
}, }, ], - "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. + "containerAnalysisAttestationProjects": [ # Optional. The projects where attestations are stored as Container Analysis Occurrences, in the format `projects/[PROJECT_ID]`. Only one attestation needs to successfully verify an image for this check to pass, so a single verified attestation found in any of `container_analysis_attestation_projects` is sufficient for the check to pass. A project ID must be used, not a project number. When fetching Occurrences from Container Analysis, only `AttestationOccurrence` kinds are considered. In the future, additional Occurrence kinds may be added to the query. Maximum number of `container_analysis_attestation_projects` allowed in each `SimpleSigningAttestationCheck` is 10. "A String", ], }, diff --git a/docs/dyn/calendar_v3.events.html b/docs/dyn/calendar_v3.events.html index c94835b0ffb..f0fbbf739b9 100644 --- a/docs/dyn/calendar_v3.events.html +++ b/docs/dyn/calendar_v3.events.html @@ -103,7 +103,7 @@Instance Methods
Retrieves the next page of results.
-
move(calendarId, eventId, destination, sendNotifications=None, sendUpdates=None)
Moves an event to another calendar, i.e. changes an event's organizer. Note that only default events can be moved; outOfOffice, focusTime and workingLocation events cannot be moved.
+Moves an event to another calendar, i.e. changes an event's organizer. Note that only default events can be moved; outOfOffice, focusTime, workingLocation and fromGmail events cannot be moved.
Updates an event. This method supports patch semantics.
@@ -312,6 +312,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -607,6 +608,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -897,6 +899,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -1191,6 +1194,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -1490,6 +1494,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -1815,6 +1820,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -1975,6 +1981,7 @@Method Details
Allowed values default - Regular events. focusTime - Focus time events. + fromGmail - Events from Gmail. outOfOffice - Out of office events. workingLocation - Working location events. iCalUID: string, Specifies an event ID in the iCalendar format to be provided in the response. Optional. Use this if you want to search for an event by its iCalendar ID. @@ -2204,6 +2211,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -2355,7 +2363,7 @@Method Details
+move(calendarId, eventId, destination, sendNotifications=None, sendUpdates=None)
-Moves an event to another calendar, i.e. changes an event's organizer. Note that only default events can be moved; outOfOffice, focusTime and workingLocation events cannot be moved. +@@ -630,6 +637,7 @@Moves an event to another calendar, i.e. changes an event's organizer. Note that only default events can be moved; outOfOffice, focusTime, workingLocation and fromGmail events cannot be moved. Args: calendarId: string, Calendar identifier of the source calendar where the event currently is on. (required) @@ -2531,6 +2539,7 @@@@ -630,6 +637,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -2826,6 +2835,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -3126,6 +3136,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -3430,6 +3441,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -3725,6 +3737,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -4025,6 +4038,7 @@Method Details
# - "outOfOffice" - An out-of-office event. # - "focusTime" - A focus-time event. # - "workingLocation" - A working location event. + # - "fromGmail" - An event from Gmail. This type of event cannot be created. "extendedProperties": { # Extended properties of the event. "private": { # Properties that are private to the copy of the event that appears on this calendar. "a_key": "A String", # The name of the private property and the corresponding value. @@ -4181,6 +4195,7 @@Method Details
Allowed values default - Regular events. focusTime - Focus time events. + fromGmail - Events from Gmail. outOfOffice - Out of office events. workingLocation - Working location events. iCalUID: string, Specifies an event ID in the iCalendar format to be provided in the response. Optional. Use this if you want to search for an event by its iCalendar ID. diff --git a/docs/dyn/chromemanagement_v1.customers.telemetry.devices.html b/docs/dyn/chromemanagement_v1.customers.telemetry.devices.html index bb5e90d277e..680cdc4f1e5 100644 --- a/docs/dyn/chromemanagement_v1.customers.telemetry.devices.html +++ b/docs/dyn/chromemanagement_v1.customers.telemetry.devices.html @@ -108,6 +108,19 @@Method Details
An object of the form: { # Telemetry data collected from a managed device. * Granular permission needed: TELEMETRY_API_DEVICE + "appReport": [ # Output only. App reports collected periodically sorted in a decreasing order of report_time. + { # App report. + "reportTime": "A String", # Timestamp when the report was collected. + "usageData": [ # App usage data. + { # App usage data. + "appId": "A String", # App id. + "appInstanceId": "A String", # Application instance id. This will be unique per window/instance. + "appType": "A String", # Type of app. + "runningDuration": "A String", # App foreground running time. + }, + ], + }, + ], "audioStatusReport": [ # Output only. Audio reports collected periodically sorted in a decreasing order of report_time. { # Status data for storage. * This field is telemetry information and this will change over time as the device is utilized. * Data for this field is controlled via policy: [ReportDeviceAudioStatus](https://chromeenterprise.google/policies/#ReportDeviceAudioStatus) * Data Collection Frequency: 10 minutes * Default Data Reporting Frequency: 3 hours - Policy Controlled: Yes * Cache: If the device is offline, the collected data is stored locally, and will be reported when the device is next online: No * Reported for affiliated users only: N/A * Granular permission needed: TELEMETRY_API_AUDIO_REPORT "inputDevice": "A String", # Output only. Active input device's name. @@ -413,6 +426,19 @@Method Details
{ "devices": [ # Telemetry devices returned in the response. { # Telemetry data collected from a managed device. * Granular permission needed: TELEMETRY_API_DEVICE + "appReport": [ # Output only. App reports collected periodically sorted in a decreasing order of report_time. + { # App report. + "reportTime": "A String", # Timestamp when the report was collected. + "usageData": [ # App usage data. + { # App usage data. + "appId": "A String", # App id. + "appInstanceId": "A String", # Application instance id. This will be unique per window/instance. + "appType": "A String", # Type of app. + "runningDuration": "A String", # App foreground running time. + }, + ], + }, + ], "audioStatusReport": [ # Output only. Audio reports collected periodically sorted in a decreasing order of report_time. { # Status data for storage. * This field is telemetry information and this will change over time as the device is utilized. * Data for this field is controlled via policy: [ReportDeviceAudioStatus](https://chromeenterprise.google/policies/#ReportDeviceAudioStatus) * Data Collection Frequency: 10 minutes * Default Data Reporting Frequency: 3 hours - Policy Controlled: Yes * Cache: If the device is offline, the collected data is stored locally, and will be reported when the device is next online: No * Reported for affiliated users only: N/A * Granular permission needed: TELEMETRY_API_AUDIO_REPORT "inputDevice": "A String", # Output only. Active input device's name. diff --git a/docs/dyn/chromemanagement_v1.customers.telemetry.users.html b/docs/dyn/chromemanagement_v1.customers.telemetry.users.html index 20d1cf0acf6..60ac3d0e83b 100644 --- a/docs/dyn/chromemanagement_v1.customers.telemetry.users.html +++ b/docs/dyn/chromemanagement_v1.customers.telemetry.users.html @@ -113,6 +113,19 @@Method Details
"orgUnitId": "A String", # Organization unit of the user. "userDevice": [ # Telemetry data collected from a managed user and device. { # Telemetry data collected for a managed user and device. * Granular permission needed: TELEMETRY_API_DEVICE + "appReport": [ # Output only. App reports collected periodically sorted in a decreasing order of report_time. + { # App report. + "reportTime": "A String", # Timestamp when the report was collected. + "usageData": [ # App usage data. + { # App usage data. + "appId": "A String", # App id. + "appInstanceId": "A String", # Application instance id. This will be unique per window/instance. + "appType": "A String", # Type of app. + "runningDuration": "A String", # App foreground running time. + }, + ], + }, + ], "audioStatusReport": [ # Output only. Audio reports collected periodically sorted in a decreasing order of report_time. { # Status data for storage. * This field is telemetry information and this will change over time as the device is utilized. * Data for this field is controlled via policy: [ReportDeviceAudioStatus](https://chromeenterprise.google/policies/#ReportDeviceAudioStatus) * Data Collection Frequency: 10 minutes * Default Data Reporting Frequency: 3 hours - Policy Controlled: Yes * Cache: If the device is offline, the collected data is stored locally, and will be reported when the device is next online: No * Reported for affiliated users only: N/A * Granular permission needed: TELEMETRY_API_AUDIO_REPORT "inputDevice": "A String", # Output only. Active input device's name. @@ -190,6 +203,19 @@Method Details
"orgUnitId": "A String", # Organization unit of the user. "userDevice": [ # Telemetry data collected from a managed user and device. { # Telemetry data collected for a managed user and device. * Granular permission needed: TELEMETRY_API_DEVICE + "appReport": [ # Output only. App reports collected periodically sorted in a decreasing order of report_time. + { # App report. + "reportTime": "A String", # Timestamp when the report was collected. + "usageData": [ # App usage data. + { # App usage data. + "appId": "A String", # App id. + "appInstanceId": "A String", # Application instance id. This will be unique per window/instance. + "appType": "A String", # Type of app. + "runningDuration": "A String", # App foreground running time. + }, + ], + }, + ], "audioStatusReport": [ # Output only. Audio reports collected periodically sorted in a decreasing order of report_time. { # Status data for storage. * This field is telemetry information and this will change over time as the device is utilized. * Data for this field is controlled via policy: [ReportDeviceAudioStatus](https://chromeenterprise.google/policies/#ReportDeviceAudioStatus) * Data Collection Frequency: 10 minutes * Default Data Reporting Frequency: 3 hours - Policy Controlled: Yes * Cache: If the device is offline, the collected data is stored locally, and will be reported when the device is next online: No * Reported for affiliated users only: N/A * Granular permission needed: TELEMETRY_API_AUDIO_REPORT "inputDevice": "A String", # Output only. Active input device's name. diff --git a/docs/dyn/cloudbuild_v2.projects.locations.connections.html b/docs/dyn/cloudbuild_v2.projects.locations.connections.html index 037f62f3642..57dee575083 100644 --- a/docs/dyn/cloudbuild_v2.projects.locations.connections.html +++ b/docs/dyn/cloudbuild_v2.projects.locations.connections.html @@ -154,7 +154,7 @@Method Details
"userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. }, - "hostUri": "A String", # Optional. The URI of the Bitbucket Data Center instance or cluster this connection is for. + "hostUri": "A String", # Required. The URI of the Bitbucket Data Center instance or cluster this connection is for. "readAuthorizerCredential": { # Represents a personal access token that authorized the Connection, and associated metadata. # Required. A http access token with the `REPO_READ` access. "userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. @@ -367,7 +367,7 @@Method Details
"userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. }, - "hostUri": "A String", # Optional. The URI of the Bitbucket Data Center instance or cluster this connection is for. + "hostUri": "A String", # Required. The URI of the Bitbucket Data Center instance or cluster this connection is for. "readAuthorizerCredential": { # Represents a personal access token that authorized the Connection, and associated metadata. # Required. A http access token with the `REPO_READ` access. "userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. @@ -518,7 +518,7 @@Method Details
"userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. }, - "hostUri": "A String", # Optional. The URI of the Bitbucket Data Center instance or cluster this connection is for. + "hostUri": "A String", # Required. The URI of the Bitbucket Data Center instance or cluster this connection is for. "readAuthorizerCredential": { # Represents a personal access token that authorized the Connection, and associated metadata. # Required. A http access token with the `REPO_READ` access. "userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. @@ -629,7 +629,7 @@Method Details
"userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. }, - "hostUri": "A String", # Optional. The URI of the Bitbucket Data Center instance or cluster this connection is for. + "hostUri": "A String", # Required. The URI of the Bitbucket Data Center instance or cluster this connection is for. "readAuthorizerCredential": { # Represents a personal access token that authorized the Connection, and associated metadata. # Required. A http access token with the `REPO_READ` access. "userTokenSecretVersion": "A String", # Required. A SecretManager resource containing the user token that authorizes the Cloud Build connection. Format: `projects/*/secrets/*/versions/*`. "username": "A String", # Output only. The username associated to this token. diff --git a/docs/dyn/cloudfunctions_v1.projects.locations.functions.html b/docs/dyn/cloudfunctions_v1.projects.locations.functions.html index 2e8cc0319ff..5a3b311fea1 100644 --- a/docs/dyn/cloudfunctions_v1.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v1.projects.locations.functions.html @@ -197,7 +197,7 @@Method Details
"name": "A String", # A user-defined name of the function. Function names must be unique globally and match pattern `projects/*/locations/*/functions/*` "network": "A String", # Deprecated: use vpc_connector "onDeployUpdatePolicy": { # Security patches are only applied when a function is redeployed. - "runtimeVersion": "A String", # Output only. contains the runtime version which was used during latest function deployment. + "runtimeVersion": "A String", # Output only. Contains the runtime version which was used during latest function deployment. }, "runtime": "A String", # The runtime in which to run the function. Required when deploying a new function, optional when updating an existing function. For a complete list of possible choices, see the [`gcloud` command reference](https://cloud.google.com/sdk/gcloud/reference/functions/deploy#--runtime). "secretEnvironmentVariables": [ # Secret environment variables configuration. @@ -409,7 +409,7 @@Method Details
"name": "A String", # A user-defined name of the function. Function names must be unique globally and match pattern `projects/*/locations/*/functions/*` "network": "A String", # Deprecated: use vpc_connector "onDeployUpdatePolicy": { # Security patches are only applied when a function is redeployed. - "runtimeVersion": "A String", # Output only. contains the runtime version which was used during latest function deployment. + "runtimeVersion": "A String", # Output only. Contains the runtime version which was used during latest function deployment. }, "runtime": "A String", # The runtime in which to run the function. Required when deploying a new function, optional when updating an existing function. For a complete list of possible choices, see the [`gcloud` command reference](https://cloud.google.com/sdk/gcloud/reference/functions/deploy#--runtime). "secretEnvironmentVariables": [ # Secret environment variables configuration. @@ -557,7 +557,7 @@Method Details
"name": "A String", # A user-defined name of the function. Function names must be unique globally and match pattern `projects/*/locations/*/functions/*` "network": "A String", # Deprecated: use vpc_connector "onDeployUpdatePolicy": { # Security patches are only applied when a function is redeployed. - "runtimeVersion": "A String", # Output only. contains the runtime version which was used during latest function deployment. + "runtimeVersion": "A String", # Output only. Contains the runtime version which was used during latest function deployment. }, "runtime": "A String", # The runtime in which to run the function. Required when deploying a new function, optional when updating an existing function. For a complete list of possible choices, see the [`gcloud` command reference](https://cloud.google.com/sdk/gcloud/reference/functions/deploy#--runtime). "secretEnvironmentVariables": [ # Secret environment variables configuration. @@ -668,7 +668,7 @@Method Details
"name": "A String", # A user-defined name of the function. Function names must be unique globally and match pattern `projects/*/locations/*/functions/*` "network": "A String", # Deprecated: use vpc_connector "onDeployUpdatePolicy": { # Security patches are only applied when a function is redeployed. - "runtimeVersion": "A String", # Output only. contains the runtime version which was used during latest function deployment. + "runtimeVersion": "A String", # Output only. Contains the runtime version which was used during latest function deployment. }, "runtime": "A String", # The runtime in which to run the function. Required when deploying a new function, optional when updating an existing function. For a complete list of possible choices, see the [`gcloud` command reference](https://cloud.google.com/sdk/gcloud/reference/functions/deploy#--runtime). "secretEnvironmentVariables": [ # Secret environment variables configuration. diff --git a/docs/dyn/cloudfunctions_v2.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2.projects.locations.functions.html index 5794af50749..16f4d335eb9 100644 --- a/docs/dyn/cloudfunctions_v2.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2.projects.locations.functions.html @@ -252,6 +252,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -268,6 +269,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -303,6 +305,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -379,6 +382,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -395,6 +399,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -421,6 +426,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -580,6 +586,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, "uploadUrl": "A String", # The generated Google Cloud Storage signed URL that should be used for a function source code upload. The uploaded file should be a zip archive which contains a function. }Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -646,6 +654,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -681,6 +690,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -757,6 +767,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -773,6 +784,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -799,6 +811,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -940,6 +953,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -956,6 +970,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -991,6 +1006,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1067,6 +1083,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1083,6 +1100,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1109,6 +1127,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1211,6 +1230,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1227,6 +1247,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1262,6 +1283,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1338,6 +1360,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1354,6 +1377,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1380,6 +1404,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, diff --git a/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html index deb195ce3b3..0d9d5ad849d 100644 --- a/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2alpha.projects.locations.functions.html @@ -252,6 +252,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -268,6 +269,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -303,6 +305,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -379,6 +382,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -395,6 +399,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -421,6 +426,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -580,6 +586,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, "uploadUrl": "A String", # The generated Google Cloud Storage signed URL that should be used for a function source code upload. The uploaded file should be a zip archive which contains a function. }Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -646,6 +654,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -681,6 +690,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -757,6 +767,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -773,6 +784,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -799,6 +811,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -940,6 +953,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -956,6 +970,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -991,6 +1006,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1067,6 +1083,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1083,6 +1100,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1109,6 +1127,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1211,6 +1230,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1227,6 +1247,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1262,6 +1283,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1338,6 +1360,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1354,6 +1377,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1380,6 +1404,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, diff --git a/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html b/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html index ed64b11b3f4..c0fa2fe4342 100644 --- a/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html +++ b/docs/dyn/cloudfunctions_v2beta.projects.locations.functions.html @@ -252,6 +252,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -268,6 +269,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -303,6 +305,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -379,6 +382,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -395,6 +399,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -421,6 +426,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -580,6 +586,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, "uploadUrl": "A String", # The generated Google Cloud Storage signed URL that should be used for a function source code upload. The uploaded file should be a zip archive which contains a function. } @@ -630,6 +637,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -646,6 +654,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -681,6 +690,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -757,6 +767,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -773,6 +784,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -799,6 +811,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -940,6 +953,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -956,6 +970,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -991,6 +1006,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1067,6 +1083,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1083,6 +1100,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1109,6 +1127,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1211,6 +1230,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1227,6 +1247,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1262,6 +1283,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, @@ -1338,6 +1360,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceProvenance": { # Provenance of the source. Ways to find the original source, or verify that some source was used for this build. # Output only. A permanent fixed identifier for source. @@ -1354,6 +1377,7 @@Method Details
"bucket": "A String", # Google Cloud Storage bucket containing the source (see [Bucket Name Requirements](https://cloud.google.com/storage/docs/bucket-naming#requirements)). "generation": "A String", # Google Cloud Storage generation for the object. If the generation is omitted, the latest generation will be used. "object": "A String", # Google Cloud Storage object containing the source. This object must be a gzipped archive file (`.tar.gz`) containing source to build. + "sourceUploadUrl": "A String", # When the specified storage bucket is a 1st gen function uploard url bucket, this field should be set as the generated upload url for 1st gen deployment. }, }, "sourceToken": "A String", # An identifier for Firebase function sources. Disclaimer: This field is only supported for Firebase function deployments. @@ -1380,6 +1404,7 @@Method Details
"allTrafficOnLatestRevision": True or False, # Whether 100% of traffic is routed to the latest revision. On CreateFunction and UpdateFunction, when set to true, the revision being deployed will serve 100% of traffic, ignoring any traffic split settings, if any. On GetFunction, true will be returned if the latest revision is serving 100% of traffic. "availableCpu": "A String", # The number of CPUs used in a single container instance. Default value is calculated from available memory. Supports the same values as Cloud Run, see https://cloud.google.com/run/docs/reference/rest/v1/Container#resourcerequirements Example: "1" indicates 1 vCPU "availableMemory": "A String", # The amount of memory available for a function. Defaults to 256M. Supported units are k, M, G, Mi, Gi. If no unit is supplied the value is interpreted as bytes. See https://github.com/kubernetes/kubernetes/blob/master/staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go a full description. + "binaryAuthorizationPolicy": "A String", # Optional. The binary authorization policy to be checked when deploying the Cloud Run service. "environmentVariables": { # Environment variables that shall be available during function execution. "a_key": "A String", }, diff --git a/docs/dyn/cloudsearch_v1.query.html b/docs/dyn/cloudsearch_v1.query.html index 30b53f7fc22..1b0674d3e49 100644 --- a/docs/dyn/cloudsearch_v1.query.html +++ b/docs/dyn/cloudsearch_v1.query.html @@ -82,6 +82,9 @@Instance Methods
Close httplib2 connections.
++
+debugSearch(body=None, x__xgafv=None)
Returns Debug information for Cloud Search Query API provides the search method. **Note:** This API requires a standard end user account to execute. A service account can't perform Query API requests directly; to use a service account to perform queries, set up [Google Workspace domain-wide delegation of authority](https://developers.google.com/cloud-search/docs/guides/delegation/).
removeActivity(body=None, x__xgafv=None)
Provides functionality to remove logged activity for a user. Currently to be used only for Chat 1p clients **Note:** This API requires a standard end user account to execute. A service account can't perform Remove Activity requests directly; to use a service account to perform queries, set up [Google Workspace domain-wide delegation of authority](https://developers.google.com/cloud-search/docs/guides/delegation/).
@@ -97,6 +100,407 @@Method Details
Close httplib2 connections.++debugSearch(body=None, x__xgafv=None)
+Returns Debug information for Cloud Search Query API provides the search method. **Note:** This API requires a standard end user account to execute. A service account can't perform Query API requests directly; to use a service account to perform queries, set up [Google Workspace domain-wide delegation of authority](https://developers.google.com/cloud-search/docs/guides/delegation/). + +Args: + body: object, The request body. + The object takes the form of: + +{ # The search API request. + "contextAttributes": [ # Context attributes for the request which will be used to adjust ranking of search results. The maximum number of elements is 10. + { # A named attribute associated with an item which can be used for influencing the ranking of the item based on the context in the request. + "name": "A String", # The name of the attribute. It should not be empty. The maximum length is 32 characters. The name must start with a letter and can only contain letters (A-Z, a-z) or numbers (0-9). The name will be normalized (lower-cased) before being matched. + "values": [ # Text values of the attribute. The maximum number of elements is 10. The maximum length of an element in the array is 32 characters. The value will be normalized (lower-cased) before being matched. + "A String", + ], + }, + ], + "dataSourceRestrictions": [ # The sources to use for querying. If not specified, all data sources from the current search application are used. + { # Restriction on Datasource. + "filterOptions": [ # Filter options restricting the results. If multiple filters are present, they are grouped by object type before joining. Filters with the same object type are joined conjunctively, then the resulting expressions are joined disjunctively. The maximum number of elements is 20. NOTE: Suggest API supports only few filters at the moment: "objecttype", "type" and "mimetype". For now, schema specific filters cannot be used to filter suggestions. + { # Filter options to be applied on query. + "filter": { # A generic way of expressing filters in a query, which supports two approaches: **1. Setting a ValueFilter.** The name must match an operator_name defined in the schema for your data source. **2. Setting a CompositeFilter.** The filters are evaluated using the logical operator. The top-level operators can only be either an AND or a NOT. AND can appear only at the top-most level. OR can appear only under a top-level AND. # Generic filter to restrict the search, such as `lang:en`, `site:xyz`. + "compositeFilter": { + "logicOperator": "A String", # The logic operator of the sub filter. + "subFilters": [ # Sub filters. + # Object with schema name: Filter + ], + }, + "valueFilter": { + "operatorName": "A String", # The `operator_name` applied to the query, such as *price_greater_than*. The filter can work against both types of filters defined in the schema for your data source: 1. `operator_name`, where the query filters results by the property that matches the value. 2. `greater_than_operator_name` or `less_than_operator_name` in your schema. The query filters the results for the property values that are greater than or less than the supplied value in the query. + "value": { # Definition of a single value with generic type. # The value to be compared with. + "booleanValue": True or False, + "dateValue": { # Represents a whole calendar date, for example a date of birth. The time of day and time zone are either specified elsewhere or are not significant. The date is relative to the [Proleptic Gregorian Calendar](https://en.wikipedia.org/wiki/Proleptic_Gregorian_calendar). The date must be a valid calendar date between the year 1 and 9999. + "day": 42, # Day of month. Must be from 1 to 31 and valid for the year and month. + "month": 42, # Month of date. Must be from 1 to 12. + "year": 42, # Year of date. Must be from 1 to 9999. + }, + "doubleValue": 3.14, + "integerValue": "A String", + "stringValue": "A String", + "timestampValue": "A String", + }, + }, + }, + "objectType": "A String", # If object_type is set, only objects of that type are returned. This should correspond to the name of the object that was registered within the definition of schema. The maximum length is 256 characters. + }, + ], + "source": { # Defines sources for the suggest/search APIs. # The source of restriction. + "name": "A String", # Source name for content indexed by the Indexing API. + "predefinedSource": "A String", # Predefined content source for Google Apps. + }, + }, + ], + "facetOptions": [ + { # Specifies operators to return facet results for. There will be one FacetResult for every source_name/object_type/operator_name combination. + "integerFacetingOptions": { # Used to specify integer faceting options. # If set, describes integer faceting options for the given integer property. The corresponding integer property in the schema should be marked isFacetable. The number of buckets returned would be minimum of this and num_facet_buckets. + "integerBuckets": [ # Buckets for given integer values should be in strictly ascending order. For example, if values supplied are (1,5,10,100), the following facet buckets will be formed {<1, [1,5), [5-10), [10-100), >=100}. + "A String", + ], + }, + "numFacetBuckets": 42, # Maximum number of facet buckets that should be returned for this facet. Defaults to 10. Maximum value is 100. + "objectType": "A String", # If object_type is set, only those objects of that type will be used to compute facets. If empty, then all objects will be used to compute facets. + "operatorName": "A String", # The name of the operator chosen for faceting. @see cloudsearch.SchemaPropertyOptions + "sourceName": "A String", # Source name to facet on. Format: datasources/{source_id} If empty, all data sources will be used. + }, + ], + "pageSize": 42, # Maximum number of search results to return in one page. Valid values are between 1 and 100, inclusive. Default value is 10. Minimum value is 50 when results beyond 2000 are requested. + "query": "A String", # The raw query string. See supported search operators in the [Narrow your search with operators](https://support.google.com/cloudsearch/answer/6172299) + "queryInterpretationOptions": { # Options to interpret user query. # Options to interpret the user query. + "disableNlInterpretation": True or False, # Flag to disable natural language (NL) interpretation of queries. Default is false, Set to true to disable natural language interpretation. NL interpretation only applies to predefined datasources. + "disableSupplementalResults": True or False, # Use this flag to disable supplemental results for a query. Supplemental results setting chosen at SearchApplication level will take precedence if set to True. + "enableVerbatimMode": True or False, # Enable this flag to turn off all internal optimizations like natural language (NL) interpretation of queries, supplemental result retrieval, and usage of synonyms including custom ones. Nl interpretation will be disabled if either one of the two flags is true. + }, + "requestOptions": { # Shared request options for all RPC methods. # Request options, such as the search application and user timezone. + "debugOptions": { # Shared request debug options for all cloudsearch RPC methods. # Debug options of the request + "enableDebugging": True or False, # If you are asked by Google to help with debugging, set this field. Otherwise, ignore this field. + }, + "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more information, see http://www.unicode.org/reports/tr35/#Unicode_locale_identifier. For translations. Set this field using the language set in browser or for the page. In the event that the user's language preference is known, set this field to the known user language. When specified, the documents in search results are biased towards the specified language. The Suggest API uses this field as a hint to make better third-party autocomplete predictions. + "searchApplicationId": "A String", # The ID generated when you create a search application using the [admin console](https://support.google.com/a/answer/9043922). + "timeZone": "A String", # Current user's time zone id, such as "America/Los_Angeles" or "Australia/Sydney". These IDs are defined by [Unicode Common Locale Data Repository (CLDR)](http://cldr.unicode.org/) project, and currently available in the file [timezone.xml](http://unicode.org/repos/cldr/trunk/common/bcp47/timezone.xml). This field is used to correctly interpret date and time queries. If this field is not specified, the default time zone (UTC) is used. + }, + "sortOptions": { # The options for sorting the search results + "operatorName": "A String", # The name of the operator corresponding to the field to sort on. The corresponding property must be marked as sortable. + "sortOrder": "A String", # Ascending is the default sort order + }, + "start": 42, # Starting index of the results. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Debug Search Response. + "gsrRequest": "A String", # Serialized string of GenericSearchRequest. + "gsrResponse": "A String", # Serialized string of GenericSearchResponse. + "searchResponse": { # The search API response. # Search response. + "debugInfo": { # Debugging information about the response. # Debugging information about the response. + "formattedDebugInfo": "A String", # General debug info formatted for display. + }, + "errorInfo": { # Error information about the response. # Error information about the response. + "errorMessages": [ + { # Error message per source response. + "errorMessage": "A String", + "source": { # Defines sources for the suggest/search APIs. + "name": "A String", # Source name for content indexed by the Indexing API. + "predefinedSource": "A String", # Predefined content source for Google Apps. + }, + }, + ], + }, + "facetResults": [ # Repeated facet results. + { # Source specific facet response + "buckets": [ # FacetBuckets for values in response containing at least a single result with the corresponding filter. + { # A bucket in a facet is the basic unit of operation. A bucket can comprise either a single value OR a contiguous range of values, depending on the type of the field bucketed. FacetBucket is currently used only for returning the response object. + "count": 42, # Number of results that match the bucket value. Counts are only returned for searches when count accuracy is ensured. Cloud Search does not guarantee facet counts for any query and facet counts might be present only intermittently, even for identical queries. Do not build dependencies on facet count existence; instead use facet ount percentages which are always returned. + "filter": { # A generic way of expressing filters in a query, which supports two approaches: **1. Setting a ValueFilter.** The name must match an operator_name defined in the schema for your data source. **2. Setting a CompositeFilter.** The filters are evaluated using the logical operator. The top-level operators can only be either an AND or a NOT. AND can appear only at the top-most level. OR can appear only under a top-level AND. # Filter to be passed in the search request if the corresponding bucket is selected. + "compositeFilter": { + "logicOperator": "A String", # The logic operator of the sub filter. + "subFilters": [ # Sub filters. + # Object with schema name: Filter + ], + }, + "valueFilter": { + "operatorName": "A String", # The `operator_name` applied to the query, such as *price_greater_than*. The filter can work against both types of filters defined in the schema for your data source: 1. `operator_name`, where the query filters results by the property that matches the value. 2. `greater_than_operator_name` or `less_than_operator_name` in your schema. The query filters the results for the property values that are greater than or less than the supplied value in the query. + "value": { # Definition of a single value with generic type. # The value to be compared with. + "booleanValue": True or False, + "dateValue": { # Represents a whole calendar date, for example a date of birth. The time of day and time zone are either specified elsewhere or are not significant. The date is relative to the [Proleptic Gregorian Calendar](https://en.wikipedia.org/wiki/Proleptic_Gregorian_calendar). The date must be a valid calendar date between the year 1 and 9999. + "day": 42, # Day of month. Must be from 1 to 31 and valid for the year and month. + "month": 42, # Month of date. Must be from 1 to 12. + "year": 42, # Year of date. Must be from 1 to 9999. + }, + "doubleValue": 3.14, + "integerValue": "A String", + "stringValue": "A String", + "timestampValue": "A String", + }, + }, + }, + "percentage": 42, # Percent of results that match the bucket value. The returned value is between (0-100], and is rounded down to an integer if fractional. If the value is not explicitly returned, it represents a percentage value that rounds to 0. Percentages are returned for all searches, but are an estimate. Because percentages are always returned, you should render percentages instead of counts. + "value": { # Definition of a single value with generic type. + "booleanValue": True or False, + "dateValue": { # Represents a whole calendar date, for example a date of birth. The time of day and time zone are either specified elsewhere or are not significant. The date is relative to the [Proleptic Gregorian Calendar](https://en.wikipedia.org/wiki/Proleptic_Gregorian_calendar). The date must be a valid calendar date between the year 1 and 9999. + "day": 42, # Day of month. Must be from 1 to 31 and valid for the year and month. + "month": 42, # Month of date. Must be from 1 to 12. + "year": 42, # Year of date. Must be from 1 to 9999. + }, + "doubleValue": 3.14, + "integerValue": "A String", + "stringValue": "A String", + "timestampValue": "A String", + }, + }, + ], + "objectType": "A String", # Object type for which facet results are returned. Can be empty. + "operatorName": "A String", # The name of the operator chosen for faceting. @see cloudsearch.SchemaPropertyOptions + "sourceName": "A String", # Source name for which facet results are returned. Will not be empty. + }, + ], + "hasMoreResults": True or False, # Whether there are more search results matching the query. + "queryInterpretation": { # Query interpretation result for user query. Empty if query interpretation is disabled. + "interpretationType": "A String", + "interpretedQuery": "A String", # The interpretation of the query used in search. For example, queries with natural language intent like "email from john" will be interpreted as "from:john source:mail". This field will not be filled when the reason is NOT_ENOUGH_RESULTS_FOUND_FOR_USER_QUERY. + "reason": "A String", # The reason for interpretation of the query. This field will not be UNSPECIFIED if the interpretation type is not NONE. + }, + "resultCountEstimate": "A String", # The estimated result count for this query. + "resultCountExact": "A String", # The exact result count for this query. + "resultCounts": { # Result count information # Expanded result count information. + "sourceResultCounts": [ # Result count information for each source with results. + { # Per source result count information. + "hasMoreResults": True or False, # Whether there are more search results for this source. + "resultCountEstimate": "A String", # The estimated result count for this source. + "resultCountExact": "A String", # The exact result count for this source. + "source": { # Defines sources for the suggest/search APIs. # The source the result count information is associated with. + "name": "A String", # Source name for content indexed by the Indexing API. + "predefinedSource": "A String", # Predefined content source for Google Apps. + }, + }, + ], + }, + "results": [ # Results from a search query. + { # Results containing indexed information for a document. + "clusteredResults": [ # If source is clustered, provide list of clustered results. There will only be one level of clustered results. If current source is not enabled for clustering, this field will be empty. + # Object with schema name: SearchResult + ], + "debugInfo": { # Debugging information about the result. # Debugging information about this search result. + "formattedDebugInfo": "A String", # General debug info formatted for display. + }, + "metadata": { # Metadata of a matched search result. # Metadata of the search result. + "createTime": "A String", # The creation time for this document or object in the search result. + "displayOptions": { # Options that specify how to display a structured data search result. + "metalines": [ # The metalines content to be displayed with the result. + { # The collection of fields that make up a displayed line + "fields": [ + { # Display Fields for Search Results + "label": "A String", # The display label for the property. + "operatorName": "A String", # The operator name of the property. + "property": { # A typed name-value pair for structured data. The type of the value should be the same as the registered type for the `name` property in the object definition of `objectType`. # The name value pair for the property. + "booleanValue": True or False, + "dateValues": { # List of date values. + "values": [ + { # Represents a whole calendar date, for example a date of birth. The time of day and time zone are either specified elsewhere or are not significant. The date is relative to the [Proleptic Gregorian Calendar](https://en.wikipedia.org/wiki/Proleptic_Gregorian_calendar). The date must be a valid calendar date between the year 1 and 9999. + "day": 42, # Day of month. Must be from 1 to 31 and valid for the year and month. + "month": 42, # Month of date. Must be from 1 to 12. + "year": 42, # Year of date. Must be from 1 to 9999. + }, + ], + }, + "doubleValues": { # List of double values. + "values": [ + 3.14, + ], + }, + "enumValues": { # List of enum values. + "values": [ # The maximum allowable length for string values is 32 characters. + "A String", + ], + }, + "htmlValues": { # List of html values. + "values": [ # The maximum allowable length for html values is 2048 characters. + "A String", + ], + }, + "integerValues": { # List of integer values. + "values": [ + "A String", + ], + }, + "name": "A String", # The name of the property. This name should correspond to the name of the property that was registered for object definition in the schema. The maximum allowable length for this property is 256 characters. + "objectValues": { # List of object values. + "values": [ + # Object with schema name: StructuredDataObject + ], + }, + "textValues": { # List of text values. + "values": [ # The maximum allowable length for text values is 2048 characters. + "A String", + ], + }, + "timestampValues": { # List of timestamp values. + "values": [ + "A String", + ], + }, + }, + }, + ], + }, + ], + "objectTypeLabel": "A String", # The display label for the object. + }, + "fields": [ # Indexed fields in structured data, returned as a generic named property. + { # A typed name-value pair for structured data. The type of the value should be the same as the registered type for the `name` property in the object definition of `objectType`. + "booleanValue": True or False, + "dateValues": { # List of date values. + "values": [ + { # Represents a whole calendar date, for example a date of birth. The time of day and time zone are either specified elsewhere or are not significant. The date is relative to the [Proleptic Gregorian Calendar](https://en.wikipedia.org/wiki/Proleptic_Gregorian_calendar). The date must be a valid calendar date between the year 1 and 9999. + "day": 42, # Day of month. Must be from 1 to 31 and valid for the year and month. + "month": 42, # Month of date. Must be from 1 to 12. + "year": 42, # Year of date. Must be from 1 to 9999. + }, + ], + }, + "doubleValues": { # List of double values. + "values": [ + 3.14, + ], + }, + "enumValues": { # List of enum values. + "values": [ # The maximum allowable length for string values is 32 characters. + "A String", + ], + }, + "htmlValues": { # List of html values. + "values": [ # The maximum allowable length for html values is 2048 characters. + "A String", + ], + }, + "integerValues": { # List of integer values. + "values": [ + "A String", + ], + }, + "name": "A String", # The name of the property. This name should correspond to the name of the property that was registered for object definition in the schema. The maximum allowable length for this property is 256 characters. + "objectValues": { # List of object values. + "values": [ + # Object with schema name: StructuredDataObject + ], + }, + "textValues": { # List of text values. + "values": [ # The maximum allowable length for text values is 2048 characters. + "A String", + ], + }, + "timestampValues": { # List of timestamp values. + "values": [ + "A String", + ], + }, + }, + ], + "mimeType": "A String", # Mime type of the search result. + "objectType": "A String", # Object type of the search result. + "owner": { # Object to represent a person. # Owner (usually creator) of the document or object of the search result. + "emailAddresses": [ # The person's email addresses + { # A person's email address. + "customType": "A String", # If the value of type is custom, this property contains the custom type string. + "emailAddress": "A String", # The email address. + "emailUrl": "A String", # The URL to send email. + "primary": True or False, # Indicates if this is the user's primary email. Only one entry can be marked as primary. + "type": "A String", # The type of the email account. Acceptable values are: "custom", "home", "other", "work". + }, + ], + "name": "A String", # The resource name of the person to provide information about. See [`People.get`](https://developers.google.com/people/api/rest/v1/people/get) from the Google People API. + "obfuscatedId": "A String", # Obfuscated ID of a person. + "personNames": [ # The person's name + { # A person's name. + "displayName": "A String", # The read-only display name formatted according to the locale specified by the viewer's account or the `Accept-Language` HTTP header. + }, + ], + "phoneNumbers": [ # The person's phone numbers + { # A person's Phone Number + "phoneNumber": "A String", # The phone number of the person. + "type": "A String", + }, + ], + "photos": [ # A person's read-only photo. A picture shown next to the person's name to help others recognize the person in search results. + { # A person's photo. + "url": "A String", # The URL of the photo. + }, + ], + }, + "source": { # Defines sources for the suggest/search APIs. # The named source for the result, such as Gmail. + "name": "A String", # Source name for content indexed by the Indexing API. + "predefinedSource": "A String", # Predefined content source for Google Apps. + }, + "thumbnailUrl": "A String", # The thumbnail URL of the result. + "updateTime": "A String", # The last modified date for the object in the search result. If not set in the item, the value returned here is empty. When `updateTime` is used for calculating freshness and is not set, this value defaults to 2 years from the current time. + }, + "snippet": { # Snippet of the search result, which summarizes the content of the resulting page. # The concatenation of all snippets (summaries) available for this result. + "matchRanges": [ # The matched ranges in the snippet. + { # Matched range of a snippet [start, end). + "end": 42, # End of the match in the snippet. + "start": 42, # Starting position of the match in the snippet. + }, + ], + "snippet": "A String", # The snippet of the document. May contain escaped HTML character that should be unescaped prior to rendering. + }, + "title": "A String", # Title of the search result. + "url": "A String", # The URL of the search result. The URL contains a Google redirect to the actual item. This URL is signed and shouldn't be changed. + }, + ], + "spellResults": [ # Suggested spelling for the query. + { + "suggestedQuery": "A String", # The suggested spelling of the query. + "suggestedQueryHtml": { # IMPORTANT: It is unsafe to accept this message from an untrusted source, since it's trivial for an attacker to forge serialized messages that don't fulfill the type's safety contract -- for example, it could contain attacker controlled script. A system which receives a SafeHtmlProto implicitly trusts the producer of the SafeHtmlProto. So, it's generally safe to return this message in RPC responses, but generally unsafe to accept it in RPC requests. # The sanitized HTML representing the spell corrected query that can be used in the UI. This usually has language-specific tags to mark up parts of the query that are spell checked. + "privateDoNotAccessOrElseSafeHtmlWrappedValue": "A String", # IMPORTANT: Never set or read this field, even from tests, it is private. See documentation at the top of .proto file for programming language packages with which to create or read this message. + }, + "suggestionType": "A String", # Suggestion triggered for the current query. + }, + ], + "structuredResults": [ # Structured results for the user query. These results are not counted against the page_size. + { # Structured results that are returned as part of search request. + "person": { # Object to represent a person. # Representation of a person + "emailAddresses": [ # The person's email addresses + { # A person's email address. + "customType": "A String", # If the value of type is custom, this property contains the custom type string. + "emailAddress": "A String", # The email address. + "emailUrl": "A String", # The URL to send email. + "primary": True or False, # Indicates if this is the user's primary email. Only one entry can be marked as primary. + "type": "A String", # The type of the email account. Acceptable values are: "custom", "home", "other", "work". + }, + ], + "name": "A String", # The resource name of the person to provide information about. See [`People.get`](https://developers.google.com/people/api/rest/v1/people/get) from the Google People API. + "obfuscatedId": "A String", # Obfuscated ID of a person. + "personNames": [ # The person's name + { # A person's name. + "displayName": "A String", # The read-only display name formatted according to the locale specified by the viewer's account or the `Accept-Language` HTTP header. + }, + ], + "phoneNumbers": [ # The person's phone numbers + { # A person's Phone Number + "phoneNumber": "A String", # The phone number of the person. + "type": "A String", + }, + ], + "photos": [ # A person's read-only photo. A picture shown next to the person's name to help others recognize the person in search results. + { # A person's photo. + "url": "A String", # The URL of the photo. + }, + ], + }, + }, + ], + }, +}+removeActivity(body=None, x__xgafv=None)
Provides functionality to remove logged activity for a user. Currently to be used only for Chat 1p clients **Note:** This API requires a standard end user account to execute. A service account can't perform Remove Activity requests directly; to use a service account to perform queries, set up [Google Workspace domain-wide delegation of authority](https://developers.google.com/cloud-search/docs/guides/delegation/). diff --git a/docs/dyn/compute_alpha.networks.html b/docs/dyn/compute_alpha.networks.html index 6f54ae6cf88..e773ccef53c 100644 --- a/docs/dyn/compute_alpha.networks.html +++ b/docs/dyn/compute_alpha.networks.html @@ -443,6 +443,7 @@@@ -159,6 +163,7 @@Method Details
"name": "A String", # Name of the resource. Provided by the client when the resource is created. The name must be 1-63 characters long, and comply with RFC1035. Specifically, the name must be 1-63 characters long and match the regular expression `[a-z]([-a-z0-9]*[a-z0-9])?`. The first character must be a lowercase letter, and all following characters (except for the last character) must be a dash, lowercase letter, or digit. The last character must be a lowercase letter or digit. "networkFirewallPolicyEnforcementOrder": "A String", # The network firewall policy enforcement order. Can be either AFTER_CLASSIC_FIREWALL or BEFORE_CLASSIC_FIREWALL. Defaults to AFTER_CLASSIC_FIREWALL if the field is not specified. "networkPlacement": "A String", # A full or partial URL of the network placement to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkPlacements/{network_placement_name} - projects/{project_id}/global/networkPlacements/{network_placement_name} + "networkProfile": "A String", # A full or partial URL of the network profile to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkProfiles/{network_profile_name} - projects/{project_id}/global/networkProfiles/{network_profile_name} "peerings": [ # [Output Only] A list of network peerings for the resource. { # A network peering attached to a network resource. The message includes the peering name, peer network, peering state, and a flag indicating whether Google Compute Engine should automatically create routes for the peering. "advertisePeerSubnetsViaRouters": True or False, # Whether Cloud Routers in this network can automatically advertise subnets from the peer network. @@ -834,6 +835,7 @@Method Details
"name": "A String", # Name of the resource. Provided by the client when the resource is created. The name must be 1-63 characters long, and comply with RFC1035. Specifically, the name must be 1-63 characters long and match the regular expression `[a-z]([-a-z0-9]*[a-z0-9])?`. The first character must be a lowercase letter, and all following characters (except for the last character) must be a dash, lowercase letter, or digit. The last character must be a lowercase letter or digit. "networkFirewallPolicyEnforcementOrder": "A String", # The network firewall policy enforcement order. Can be either AFTER_CLASSIC_FIREWALL or BEFORE_CLASSIC_FIREWALL. Defaults to AFTER_CLASSIC_FIREWALL if the field is not specified. "networkPlacement": "A String", # A full or partial URL of the network placement to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkPlacements/{network_placement_name} - projects/{project_id}/global/networkPlacements/{network_placement_name} + "networkProfile": "A String", # A full or partial URL of the network profile to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkProfiles/{network_profile_name} - projects/{project_id}/global/networkProfiles/{network_profile_name} "peerings": [ # [Output Only] A list of network peerings for the resource. { # A network peering attached to a network resource. The message includes the peering name, peer network, peering state, and a flag indicating whether Google Compute Engine should automatically create routes for the peering. "advertisePeerSubnetsViaRouters": True or False, # Whether Cloud Routers in this network can automatically advertise subnets from the peer network. @@ -1020,6 +1022,7 @@Method Details
"name": "A String", # Name of the resource. Provided by the client when the resource is created. The name must be 1-63 characters long, and comply with RFC1035. Specifically, the name must be 1-63 characters long and match the regular expression `[a-z]([-a-z0-9]*[a-z0-9])?`. The first character must be a lowercase letter, and all following characters (except for the last character) must be a dash, lowercase letter, or digit. The last character must be a lowercase letter or digit. "networkFirewallPolicyEnforcementOrder": "A String", # The network firewall policy enforcement order. Can be either AFTER_CLASSIC_FIREWALL or BEFORE_CLASSIC_FIREWALL. Defaults to AFTER_CLASSIC_FIREWALL if the field is not specified. "networkPlacement": "A String", # A full or partial URL of the network placement to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkPlacements/{network_placement_name} - projects/{project_id}/global/networkPlacements/{network_placement_name} + "networkProfile": "A String", # A full or partial URL of the network profile to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkProfiles/{network_profile_name} - projects/{project_id}/global/networkProfiles/{network_profile_name} "peerings": [ # [Output Only] A list of network peerings for the resource. { # A network peering attached to a network resource. The message includes the peering name, peer network, peering state, and a flag indicating whether Google Compute Engine should automatically create routes for the peering. "advertisePeerSubnetsViaRouters": True or False, # Whether Cloud Routers in this network can automatically advertise subnets from the peer network. @@ -1301,6 +1304,7 @@Method Details
"name": "A String", # Name of the resource. Provided by the client when the resource is created. The name must be 1-63 characters long, and comply with RFC1035. Specifically, the name must be 1-63 characters long and match the regular expression `[a-z]([-a-z0-9]*[a-z0-9])?`. The first character must be a lowercase letter, and all following characters (except for the last character) must be a dash, lowercase letter, or digit. The last character must be a lowercase letter or digit. "networkFirewallPolicyEnforcementOrder": "A String", # The network firewall policy enforcement order. Can be either AFTER_CLASSIC_FIREWALL or BEFORE_CLASSIC_FIREWALL. Defaults to AFTER_CLASSIC_FIREWALL if the field is not specified. "networkPlacement": "A String", # A full or partial URL of the network placement to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkPlacements/{network_placement_name} - projects/{project_id}/global/networkPlacements/{network_placement_name} + "networkProfile": "A String", # A full or partial URL of the network profile to apply to this network. This field can be set only at resource creation time. For example, the following are valid URLs: - https://www.googleapis.com/compute/alpha/projects/{project_id}/global/networkProfiles/{network_profile_name} - projects/{project_id}/global/networkProfiles/{network_profile_name} "peerings": [ # [Output Only] A list of network peerings for the resource. { # A network peering attached to a network resource. The message includes the peering name, peer network, peering state, and a flag indicating whether Google Compute Engine should automatically create routes for the peering. "advertisePeerSubnetsViaRouters": True or False, # Whether Cloud Routers in this network can automatically advertise subnets from the peer network. diff --git a/docs/dyn/compute_alpha.regionZones.html b/docs/dyn/compute_alpha.regionZones.html index 871055f89eb..d737f9c6c1a 100644 --- a/docs/dyn/compute_alpha.regionZones.html +++ b/docs/dyn/compute_alpha.regionZones.html @@ -112,7 +112,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/compute_alpha.zones.html b/docs/dyn/compute_alpha.zones.html index b788aa4da59..57c0b5b3a45 100644 --- a/docs/dyn/compute_alpha.zones.html +++ b/docs/dyn/compute_alpha.zones.html @@ -107,7 +107,7 @@Method Details
Returns: An object of the form: - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], @@ -158,7 +158,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/compute_beta.instanceGroupManagers.html b/docs/dyn/compute_beta.instanceGroupManagers.html index 7d35f770a52..94b629af2ff 100644 --- a/docs/dyn/compute_beta.instanceGroupManagers.html +++ b/docs/dyn/compute_beta.instanceGroupManagers.html @@ -400,6 +400,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -422,6 +426,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1329,6 +1335,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1351,6 +1361,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1506,6 +1518,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1528,6 +1544,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1812,6 +1830,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1834,6 +1856,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -2135,6 +2159,7 @@Method Details
}, "propertiesFromFlexibilityPolicy": { # [Output Only] Instance properties selected for this instance resulting from InstanceFlexibilityPolicy. "machineType": "A String", # The machine type to be used for this instance. + "provisioningModel": "A String", # The provisioning model to be used for this instance. }, "targetStatus": "A String", # [Output Only] The eventual status of the instance. The instance group manager will not be identified as stable till each managed instance reaches its targetStatus. "version": { # [Output Only] Intended version of this instance. @@ -2336,6 +2361,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -2358,6 +2387,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -4165,6 +4196,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -4187,6 +4222,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. diff --git a/docs/dyn/compute_beta.instanceTemplates.html b/docs/dyn/compute_beta.instanceTemplates.html index 96610855977..6f3dd5e6efc 100644 --- a/docs/dyn/compute_beta.instanceTemplates.html +++ b/docs/dyn/compute_beta.instanceTemplates.html @@ -205,7 +205,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -213,8 +213,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -256,7 +256,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -699,7 +699,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -707,8 +707,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -750,7 +750,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -1112,7 +1112,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1120,8 +1120,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1163,7 +1163,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -1559,7 +1559,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1567,8 +1567,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1610,7 +1610,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", diff --git a/docs/dyn/compute_beta.instances.html b/docs/dyn/compute_beta.instances.html index 74e93b8d26b..26bfa07cfed 100644 --- a/docs/dyn/compute_beta.instances.html +++ b/docs/dyn/compute_beta.instances.html @@ -616,7 +616,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -624,8 +624,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -667,7 +667,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -995,7 +995,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1003,8 +1003,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1046,7 +1046,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -1248,7 +1248,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1256,8 +1256,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1299,7 +1299,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -2076,7 +2076,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -2084,8 +2084,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -2127,7 +2127,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -3016,7 +3016,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3024,8 +3024,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3067,7 +3067,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -3503,7 +3503,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3511,8 +3511,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3554,7 +3554,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -7407,7 +7407,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -7415,8 +7415,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -7458,7 +7458,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", diff --git a/docs/dyn/compute_beta.machineImages.html b/docs/dyn/compute_beta.machineImages.html index 8307e0bb8cc..77161ccf8c4 100644 --- a/docs/dyn/compute_beta.machineImages.html +++ b/docs/dyn/compute_beta.machineImages.html @@ -316,7 +316,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -324,8 +324,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -367,7 +367,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -912,7 +912,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -920,8 +920,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -963,7 +963,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -1538,7 +1538,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1546,8 +1546,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1589,7 +1589,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", diff --git a/docs/dyn/compute_beta.networks.html b/docs/dyn/compute_beta.networks.html index 21be92bf2fe..cadc7eb484c 100644 --- a/docs/dyn/compute_beta.networks.html +++ b/docs/dyn/compute_beta.networks.html @@ -106,7 +106,7 @@Instance Methods
Retrieves the next page of results.
-
patch(project, network, body=None, requestId=None, x__xgafv=None)
Patches the specified network with the data included in the request. Only the following fields can be modified: routingConfig.routingMode.
+Patches the specified network with the data included in the request. Only routingConfig can be modified.
removePeering(project, network, body=None, requestId=None, x__xgafv=None)
Removes a peering from the specified network.
@@ -1093,7 +1093,7 @@Method Details
@@ -405,6 +407,7 @@patch(project, network, body=None, requestId=None, x__xgafv=None)
-Patches the specified network with the data included in the request. Only the following fields can be modified: routingConfig.routingMode. +Patches the specified network with the data included in the request. Only routingConfig can be modified. Args: project: string, Project ID for this request. (required) diff --git a/docs/dyn/compute_beta.regionInstanceGroupManagers.html b/docs/dyn/compute_beta.regionInstanceGroupManagers.html index 2553b5b14b5..3a97cfe0e54 100644 --- a/docs/dyn/compute_beta.regionInstanceGroupManagers.html +++ b/docs/dyn/compute_beta.regionInstanceGroupManagers.html @@ -1086,6 +1086,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1108,6 +1112,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1263,6 +1269,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1285,6 +1295,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1569,6 +1581,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -1591,6 +1607,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -1892,6 +1910,7 @@Method Details
}, "propertiesFromFlexibilityPolicy": { # [Output Only] Instance properties selected for this instance resulting from InstanceFlexibilityPolicy. "machineType": "A String", # The machine type to be used for this instance. + "provisioningModel": "A String", # The provisioning model to be used for this instance. }, "targetStatus": "A String", # [Output Only] The eventual status of the instance. The instance group manager will not be identified as stable till each managed instance reaches its targetStatus. "version": { # [Output Only] Intended version of this instance. @@ -2093,6 +2112,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -2115,6 +2138,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. @@ -3922,6 +3947,10 @@Method Details
"rank": 42, # Preference of this instance selection. Lower number means higher preference. MIG will first try to create a VM based on the machine-type with lowest rank and fallback to next rank based on availability. Machine types and instance selections with the same rank have the same preference. }, }, + "provisioningModelMix": { # Provisioning model configuration used by this managed instance group to create instances. + "standardCapacityBase": 42, # The base capacity that will always use Standard VMs to avoid risk of more preemption than the minimum capacity user needs. MIG will create only Standard VMs until it reaches standard_capacity_base and only then will start using standard_capacity_percent_above_base to mix Spot with Standard VMs. + "standardCapacityPercentAboveBase": 42, # The percentage of target capacity that should use Standard VM. The remaining percentage will use Spot VMs. The percentage applies only to the capacity above standard_capacity_base. + }, }, "instanceGroup": "A String", # [Output Only] The URL of the Instance Group resource. "instanceLifecyclePolicy": { # The repair policy for this managed instance group. @@ -3944,6 +3973,8 @@Method Details
}, }, "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "serviceAccount": "A String", # The service account to be used as credentials for all operations performed by the managed instance group on instances. The service accounts needs all permissions required to create and delete instances. By default, the service account {projectNumber}@cloudservices.gserviceaccount.com is used. "standbyPolicy": { # Standby policy for stopped and suspended instances. diff --git a/docs/dyn/compute_beta.regionInstanceTemplates.html b/docs/dyn/compute_beta.regionInstanceTemplates.html index 624923c00df..611ffe9837e 100644 --- a/docs/dyn/compute_beta.regionInstanceTemplates.html +++ b/docs/dyn/compute_beta.regionInstanceTemplates.html @@ -315,7 +315,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -323,8 +323,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -366,7 +366,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -629,7 +629,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -637,8 +637,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -680,7 +680,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", @@ -1077,7 +1077,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1085,8 +1085,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1128,7 +1128,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", diff --git a/docs/dyn/compute_beta.regionInstances.html b/docs/dyn/compute_beta.regionInstances.html index f4178039776..f92c2b2bd29 100644 --- a/docs/dyn/compute_beta.regionInstances.html +++ b/docs/dyn/compute_beta.regionInstances.html @@ -158,7 +158,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -166,8 +166,8 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceInstantSnapshot": "A String", # The source instant-snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceInstantSnapshot initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: us-central1-a/instantSnapshots/my-backup If the source instant-snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -209,7 +209,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. "userLicenses": [ # [Output Only] A list of user provided licenses. It represents a list of URLs to the license resource. Unlike regular licenses, user provided licenses can be modified after the disk is created. "A String", diff --git a/docs/dyn/compute_beta.regionZones.html b/docs/dyn/compute_beta.regionZones.html index eda5fa3d064..3a92644987f 100644 --- a/docs/dyn/compute_beta.regionZones.html +++ b/docs/dyn/compute_beta.regionZones.html @@ -112,7 +112,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/compute_beta.zones.html b/docs/dyn/compute_beta.zones.html index 4bf0b6c3872..c72efc5bce1 100644 --- a/docs/dyn/compute_beta.zones.html +++ b/docs/dyn/compute_beta.zones.html @@ -107,7 +107,7 @@Method Details
Returns: An object of the form: - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], @@ -158,7 +158,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/compute_v1.instanceGroupManagers.html b/docs/dyn/compute_v1.instanceGroupManagers.html index 5b411ea82d7..ea71371d261 100644 --- a/docs/dyn/compute_v1.instanceGroupManagers.html +++ b/docs/dyn/compute_v1.instanceGroupManagers.html @@ -374,6 +374,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1270,6 +1272,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1414,6 +1418,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1687,6 +1693,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -2168,6 +2176,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. diff --git a/docs/dyn/compute_v1.instanceTemplates.html b/docs/dyn/compute_v1.instanceTemplates.html index 5910f464fcb..d84f56ec3f6 100644 --- a/docs/dyn/compute_v1.instanceTemplates.html +++ b/docs/dyn/compute_v1.instanceTemplates.html @@ -198,7 +198,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -206,7 +206,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -247,7 +247,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -648,7 +648,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -656,7 +656,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -697,7 +697,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -1022,7 +1022,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1030,7 +1030,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1071,7 +1071,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -1425,7 +1425,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1433,7 +1433,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1474,7 +1474,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], diff --git a/docs/dyn/compute_v1.instances.html b/docs/dyn/compute_v1.instances.html index 060be2b0284..42dbdca60c7 100644 --- a/docs/dyn/compute_v1.instances.html +++ b/docs/dyn/compute_v1.instances.html @@ -594,7 +594,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -602,7 +602,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -643,7 +643,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -931,7 +931,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -939,7 +939,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -980,7 +980,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. } @@ -1172,7 +1172,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1180,7 +1180,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1221,7 +1221,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -1956,7 +1956,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1964,7 +1964,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -2005,7 +2005,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -2633,7 +2633,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -2641,7 +2641,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -2682,7 +2682,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -3072,7 +3072,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3080,7 +3080,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -3121,7 +3121,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -6626,7 +6626,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -6634,7 +6634,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -6675,7 +6675,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], diff --git a/docs/dyn/compute_v1.machineImages.html b/docs/dyn/compute_v1.machineImages.html index 233d15d3323..bdef0a55730 100644 --- a/docs/dyn/compute_v1.machineImages.html +++ b/docs/dyn/compute_v1.machineImages.html @@ -309,7 +309,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -317,7 +317,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -358,7 +358,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -854,7 +854,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -862,7 +862,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -903,7 +903,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -1429,7 +1429,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1437,7 +1437,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1478,7 +1478,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], diff --git a/docs/dyn/compute_v1.networks.html b/docs/dyn/compute_v1.networks.html index 5aea0c5d43b..95253b3ede6 100644 --- a/docs/dyn/compute_v1.networks.html +++ b/docs/dyn/compute_v1.networks.html @@ -106,7 +106,7 @@Instance Methods
Retrieves the next page of results.
-
patch(project, network, body=None, requestId=None, x__xgafv=None)
Patches the specified network with the data included in the request. Only the following fields can be modified: routingConfig.routingMode.
+Patches the specified network with the data included in the request. Only routingConfig can be modified.
removePeering(project, network, body=None, requestId=None, x__xgafv=None)
Removes a peering from the specified network.
@@ -928,7 +928,7 @@Method Details
@@ -300,6 +301,7 @@patch(project, network, body=None, requestId=None, x__xgafv=None)
-Patches the specified network with the data included in the request. Only the following fields can be modified: routingConfig.routingMode. +Patches the specified network with the data included in the request. Only routingConfig can be modified. Args: project: string, Project ID for this request. (required) diff --git a/docs/dyn/compute_v1.regionInstanceGroupManagers.html b/docs/dyn/compute_v1.regionInstanceGroupManagers.html index f6469442f74..d545c66f463 100644 --- a/docs/dyn/compute_v1.regionInstanceGroupManagers.html +++ b/docs/dyn/compute_v1.regionInstanceGroupManagers.html @@ -1060,6 +1060,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1204,6 +1206,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1477,6 +1481,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. @@ -1958,6 +1964,8 @@Method Details
}, ], "region": "A String", # [Output Only] The URL of the region where the managed instance group resides (for regional resources). + "satisfiesPzi": True or False, # [Output Only] Reserved for future use. + "satisfiesPzs": True or False, # [Output Only] Reserved for future use. "selfLink": "A String", # [Output Only] The URL for this managed instance group. The server defines this URL. "statefulPolicy": { # Stateful configuration for this Instanced Group Manager "preservedState": { # Configuration of preserved resources. diff --git a/docs/dyn/compute_v1.regionInstanceTemplates.html b/docs/dyn/compute_v1.regionInstanceTemplates.html index 0ddee561aaa..5506d7bbe84 100644 --- a/docs/dyn/compute_v1.regionInstanceTemplates.html +++ b/docs/dyn/compute_v1.regionInstanceTemplates.html @@ -303,7 +303,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -311,7 +311,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -352,7 +352,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -578,7 +578,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -586,7 +586,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -627,7 +627,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], @@ -982,7 +982,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -990,7 +990,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -1031,7 +1031,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], diff --git a/docs/dyn/compute_v1.regionInstances.html b/docs/dyn/compute_v1.regionInstances.html index 7822dba3372..5295fda869e 100644 --- a/docs/dyn/compute_v1.regionInstances.html +++ b/docs/dyn/compute_v1.regionInstances.html @@ -151,7 +151,7 @@Method Details
"resourcePolicies": [ # Resource policies applied to this disk for automatic snapshot creations. Specified using the full or partial URL. For instance template, specify only the resource policy name. "A String", ], - "sourceImage": "A String", # The source image to create this disk. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. + "sourceImage": "A String", # The source image to create this disk. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. To create a disk with one of the public operating system images, specify the image by its family name. For example, specify family/debian-9 to use the latest Debian 9 image: projects/debian-cloud/global/images/family/debian-9 Alternatively, use a specific version of a public operating system image: projects/debian-cloud/global/images/debian-9-stretch-vYYYYMMDD To create a disk with a custom image that you created, specify the image name in the following format: global/images/my-custom-image You can also specify a custom image by its image family, which returns the latest version of the image in that family. Replace the image name with family/family-name: global/images/family/my-image-family If the source image is deleted later, this field will not be set. "sourceImageEncryptionKey": { # The customer-supplied encryption key of the source image. Required if the source image is protected by a customer-supplied encryption key. InstanceTemplate and InstancePropertiesPatch do not store customer-supplied encryption keys, so you cannot create disks for instances in a managed instance group if the source images are encrypted with your own keys. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -159,7 +159,7 @@Method Details
"rsaEncryptedKey": "A String", # Specifies an RFC 4648 base64 encoded, RSA-wrapped 2048-bit customer-supplied encryption key to either encrypt or decrypt this resource. You can provide either the rawKey or the rsaEncryptedKey. For example: "rsaEncryptedKey": "ieCx/NcW06PcT7Ep1X6LUTc/hLvUDYyzSZPPVCVPTVEohpeHASqC8uw5TzyO9U+Fka9JFH z0mBibXUInrC/jEk014kCK/NPjYgEMOyssZ4ZINPKxlUh2zn1bV+MCaTICrdmuSBTWlUUiFoD D6PYznLwh8ZNdaheCeZ8ewEXgFQ8V+sDroLaN3Xs3MDTXQEMMoNUXMCZEIpg9Vtp9x2oe==" The key must meet the following requirements before you can provide it to Compute Engine: 1. The key is wrapped using a RSA public key certificate provided by Google. 2. After being wrapped, the key must be encoded in RFC 4648 base64 encoding. Gets the RSA public key certificate provided by Google at: https://cloud-certs.storage.googleapis.com/google-cloud-csek-ingress.pem "sha256": "A String", # [Output only] The RFC 4648 base64 encoded SHA-256 hash of the customer-supplied encryption key that protects this resource. }, - "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required except for local SSD. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. + "sourceSnapshot": "A String", # The source snapshot to create this disk. When creating a new instance boot disk, one of initializeParams.sourceSnapshot or initializeParams.sourceImage or disks.source is required. To create a disk with a snapshot that you created, specify the snapshot name in the following format: global/snapshots/my-backup If the source snapshot is deleted later, this field will not be set. "sourceSnapshotEncryptionKey": { # The customer-supplied encryption key of the source snapshot. "kmsKeyName": "A String", # The name of the encryption key that is stored in Google Cloud KMS. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key The fully-qualifed key name may be returned for resource GET requests. For example: "kmsKeyName": "projects/kms_project_id/locations/region/keyRings/ key_region/cryptoKeys/key /cryptoKeyVersions/1 "kmsKeyServiceAccount": "A String", # The service account being used for the encryption request for the given KMS key. If absent, the Compute Engine default service account is used. For example: "kmsKeyServiceAccount": "name@project_id.iam.gserviceaccount.com/ @@ -200,7 +200,7 @@Method Details
"fileType": "A String", # The file type of source file. }, }, - "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required except for local SSD. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. + "source": "A String", # Specifies a valid partial or full URL to an existing Persistent Disk resource. When creating a new instance boot disk, one of initializeParams.sourceImage or initializeParams.sourceSnapshot or disks.source is required. If desired, you can also attach existing non-root persistent disks using this property. This field is only applicable for persistent disks. Note that for InstanceTemplate, specify the disk name for zonal disk, and the URL for regional disk. "type": "A String", # Specifies the type of the disk, either SCRATCH or PERSISTENT. If not specified, the default is PERSISTENT. }, ], diff --git a/docs/dyn/compute_v1.regionTargetHttpsProxies.html b/docs/dyn/compute_v1.regionTargetHttpsProxies.html index dd3d5a9fda2..1a2e2e0504c 100644 --- a/docs/dyn/compute_v1.regionTargetHttpsProxies.html +++ b/docs/dyn/compute_v1.regionTargetHttpsProxies.html @@ -267,6 +267,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map }Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map } @@ -462,6 +464,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map }, ], @@ -525,6 +528,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map } diff --git a/docs/dyn/compute_v1.regionZones.html b/docs/dyn/compute_v1.regionZones.html index b3301391c9a..1885d93598d 100644 --- a/docs/dyn/compute_v1.regionZones.html +++ b/docs/dyn/compute_v1.regionZones.html @@ -112,7 +112,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/compute_v1.targetHttpsProxies.html b/docs/dyn/compute_v1.targetHttpsProxies.html index 03233c7f5d8..43bf45df923 100644 --- a/docs/dyn/compute_v1.targetHttpsProxies.html +++ b/docs/dyn/compute_v1.targetHttpsProxies.html @@ -162,6 +162,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map }, ], @@ -373,6 +374,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map }Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map } @@ -566,6 +569,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map }, ], @@ -628,6 +632,7 @@Method Details
"A String", ], "sslPolicy": "A String", # URL of SslPolicy resource that will be associated with the TargetHttpsProxy resource. If not set, the TargetHttpsProxy resource has no SSL policy configured. + "tlsEarlyData": "A String", # Specifies whether TLS 1.3 0-RTT Data ("Early Data") should be accepted for this service. Early Data allows a TLS resumption handshake to include the initial application payload (a HTTP request) alongside the handshake, reducing the effective round trips to "zero". This applies to TLS 1.3 connections over TCP (HTTP/2) as well as over UDP (QUIC/h3). This can improve application performance, especially on networks where interruptions may be common, such as on mobile. Requests with Early Data will have the "Early-Data" HTTP header set on the request, with a value of "1", to allow the backend to determine whether Early Data was included. Note: TLS Early Data may allow requests to be replayed, as the data is sent to the backend before the handshake has fully completed. Applications that allow idempotent HTTP methods to make non-idempotent changes, such as a GET request updating a database, should not accept Early Data on those requests, and reject requests with the "Early-Data: 1" HTTP header by returning a HTTP 425 (Too Early) status code, in order to remain RFC compliant. The default value is DISABLED. "urlMap": "A String", # A fully-qualified or valid partial URL to the UrlMap resource that defines the mapping from URL to the BackendService. For example, the following are all valid URLs for specifying a URL map: - https://www.googleapis.compute/v1/projects/project/global/urlMaps/ url-map - projects/project/global/urlMaps/url-map - global/urlMaps/url-map } diff --git a/docs/dyn/compute_v1.zones.html b/docs/dyn/compute_v1.zones.html index b5cf6d51b5f..6deb1de2066 100644 --- a/docs/dyn/compute_v1.zones.html +++ b/docs/dyn/compute_v1.zones.html @@ -107,7 +107,7 @@Method Details
Returns: An object of the form: - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], @@ -152,7 +152,7 @@Method Details
{ # Contains a list of zone resources. "id": "A String", # [Output Only] Unique identifier for the resource; defined by the server. "items": [ # A list of Zone resources. - { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-a is located in the us-east1 region. For more information, read Regions and Zones. + { # Represents a Zone resource. A zone is a deployment area. These deployment areas are subsets of a region. For example the zone us-east1-b is located in the us-east1 region. For more information, read Regions and Zones. "availableCpuPlatforms": [ # [Output Only] Available cpu/platform selections for the zone. "A String", ], diff --git a/docs/dyn/connectors_v1.projects.locations.providers.connectors.html b/docs/dyn/connectors_v1.projects.locations.providers.connectors.html index 1c148299209..32a4f2a12d4 100644 --- a/docs/dyn/connectors_v1.projects.locations.providers.connectors.html +++ b/docs/dyn/connectors_v1.projects.locations.providers.connectors.html @@ -112,6 +112,7 @@Method Details
An object of the form: { # Connectors indicates a specific connector type, e.x. Salesforce, SAP etc. + "category": "A String", # Output only. Category of the connector. "createTime": "A String", # Output only. Created time. "description": "A String", # Output only. Description of the resource. "displayName": "A String", # Output only. Display name. @@ -134,6 +135,9 @@Method Details
}, "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage. "name": "A String", # Output only. Resource name of the Connector. Format: projects/{project}/locations/{location}/providers/{provider}/connectors/{connector} Only global location is supported for Connector resource. + "tags": [ # Output only. Tags of the connector. + "A String", + ], "updateTime": "A String", # Output only. Updated time. "webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI. }Method Details
{ # Response message for Connectors.ListConnectors. "connectors": [ # A list of connectors. { # Connectors indicates a specific connector type, e.x. Salesforce, SAP etc. + "category": "A String", # Output only. Category of the connector. "createTime": "A String", # Output only. Created time. "description": "A String", # Output only. Description of the resource. "displayName": "A String", # Output only. Display name. @@ -181,6 +186,9 @@Method Details
}, "launchStage": "A String", # Output only. Flag to mark the version indicating the launch stage. "name": "A String", # Output only. Resource name of the Connector. Format: projects/{project}/locations/{location}/providers/{provider}/connectors/{connector} Only global location is supported for Connector resource. + "tags": [ # Output only. Tags of the connector. + "A String", + ], "updateTime": "A String", # Output only. Updated time. "webAssetsLocation": "A String", # Output only. Cloud storage location of icons etc consumed by UI. }, diff --git a/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.entitieswithacls.html b/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.entitieswithacls.html new file mode 100644 index 00000000000..a6cbef26b24 --- /dev/null +++ b/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.entitieswithacls.html @@ -0,0 +1,148 @@ + + + +Connectors API . projects . locations . connections . entityTypes . entitieswithacls
+Instance Methods
++
+close()
Close httplib2 connections.
+ +Lists entity rows with ACLs of a particular entity type contained in the request. Note: 1. Currently, only max of one 'sort_by' column is supported. 2. If no 'sort_by' column is provided, the primary key of the table is used. If zero or more than one primary key is available, we default to the unpaginated list entities logic which only returns the first page. 3. The values of the 'sort_by' columns must uniquely identify an entity row, otherwise undefined behaviors may be observed during pagination. 4. Since transactions are not supported, any updates, inserts or deletes during pagination can lead to stale data being returned or other unexpected behaviors.
+ +Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +list(parent, conditions=None, gsutilUri=None, pageSize=None, pageToken=None, sortBy=None, x__xgafv=None)
+Lists entity rows with ACLs of a particular entity type contained in the request. Note: 1. Currently, only max of one 'sort_by' column is supported. 2. If no 'sort_by' column is provided, the primary key of the table is used. If zero or more than one primary key is available, we default to the unpaginated list entities logic which only returns the first page. 3. The values of the 'sort_by' columns must uniquely identify an entity row, otherwise undefined behaviors may be observed during pagination. 4. Since transactions are not supported, any updates, inserts or deletes during pagination can lead to stale data being returned or other unexpected behaviors. + +Args: + parent: string, Required. Resource name of the Entity Type. Format: projects/{project}/locations/{location}/connections/{connection}/entityTypes/{type} (required) + conditions: string, Conditions to be used when listing entities. From a proto standpoint, There are no restrictions on what can be passed using this field. The connector documentation should have information about what format of filters/conditions are supported. + gsutilUri: string, Format: gs://object_path + pageSize: integer, Number of entity rows to return. Defaults page size = 25. Max page size = 200. + pageToken: string, Page token value if available from a previous request. + sortBy: string, List of 'sort_by' columns to use when returning the results. (repeated) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response message for EntityService.ListEntitiesWithACLs + "entitiesWithAcl": [ # List containing entity rows. + { # EntityWithACL refers to a single row of an entity type with ACL information. + "acl_info": { # AclInfo has a list of readers for a resource. This is defined as per the below docs https://cloud.google.com/generative-ai-app-builder/docs/reference/rest/v1alpha/projects.locations.collections.dataStores.branches.documents#aclinfo # ACL information of the entity. + "readers": [ # A list of readers for a resource. + { # Readers is a list of principals that have read access to a resource. + "principals": [ # A list of principals that have read access to a resource. + { # Principal is a user or group that has access to a resource. + "group_id": "A String", # The group that has access to a resource. + "user_id": "A String", # The user that has access to a resource. + }, + ], + }, + ], + }, + "id": "A String", + "jsonData": "A String", # Entity data in JSON format. + }, + ], + "nextPageToken": "A String", # Next page token if more records are available. +}+++ + \ No newline at end of file diff --git a/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.html b/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.html index 85b93f52879..b54bea8be30 100644 --- a/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.html +++ b/docs/dyn/connectors_v2.projects.locations.connections.entityTypes.html @@ -79,6 +79,11 @@list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Instance Methods
Returns the entities Resource.
+ +Returns the entitieswithacls Resource.
+Close httplib2 connections.
diff --git a/docs/dyn/contactcenteraiplatform_v1alpha1.projects.locations.contactCenters.html b/docs/dyn/contactcenteraiplatform_v1alpha1.projects.locations.contactCenters.html index 42629bd445b..9e9da011b60 100644 --- a/docs/dyn/contactcenteraiplatform_v1alpha1.projects.locations.contactCenters.html +++ b/docs/dyn/contactcenteraiplatform_v1alpha1.projects.locations.contactCenters.html @@ -159,6 +159,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -169,6 +172,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -329,6 +335,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -339,6 +348,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -439,6 +451,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -449,6 +464,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -558,6 +576,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], @@ -568,6 +589,9 @@Method Details
"name": "A String", # Name of the component. "serviceAttachments": [ # Associated service attachments. { # Container for the VPC-SC networking configurations. + "allowedProjectIds": [ # The list of project ids that are allowed to send traffic to the service attachment. This field should be filled only for the ingress service attachments. + "A String", + ], "name": "A String", # The service attachment name that will be used for sending private traffic to the CCAIP tenant project. Example: "projects/${TENANT_PROJECT_ID}/regions/${REGION}/serviceAttachments/ingress-default". }, ], diff --git a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html index ccb1bb0b28d..a4bb4cd89e0 100644 --- a/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html +++ b/docs/dyn/contactcenterinsights_v1.projects.locations.conversations.html @@ -515,98 +515,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -627,6 +535,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -906,98 +818,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -1018,6 +838,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -1327,98 +1151,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -1439,6 +1171,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -1801,98 +1537,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -1913,6 +1557,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -2210,98 +1858,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -2322,6 +1878,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -2601,98 +2161,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -2713,6 +2181,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. @@ -2994,98 +2466,6 @@Method Details
"question": "A String", # The corresponding FAQ question. "source": "A String", # The knowledge document that this answer was extracted from. Format: projects/{project}/knowledgeBases/{knowledge_base}/documents/{document}. }, - "generatorSuggestionResult": { # Represents response from generators. # The generator suggestion result. - "generatorSuggestion": { # Suggestion generated using a Generator. # The suggestion generated from the Generator. - "agentCoachingSuggestion": { # Suggestion for coaching agents. # Optional. Suggestion to coach the agent. - "agentActionSuggestions": [ # Optional. Suggested actions for the agent to take. - { # Actions suggested for the agent. This is based on applicable instructions. - "agentAction": "A String", # Optional. The suggested action for the agent. - }, - ], - "applicableInstructions": [ # Optional. Instructions applicable based on the current context. - { # Agent Coaching instructions that customer can configure. - "agentAction": "A String", # Optional. The action that human agent should take. For example, "apologize for the slow shipping". If the users only want to use agent coaching for intent detection, agent_action can be empty - "condition": "A String", # Optional. The condition of the instruction. For example, "the customer wants to cancel an order". If the users want the instruction to be triggered unconditionally, the condition can be empty. - "description": "A String", # Optional. The detailed description of this instruction. - "displayName": "A String", # Optional. Display name for the instruction. - "metadata": { # Optional. Additional information attached to this instruction. - "a_key": "A String", - }, - "systemAction": "A String", # Optional. The action that system should take. For example, "call GetOrderTime with order_number={order number provided by the customer}". If the users don't have plugins or don't want to trigger plugins, the system_action can be empty - }, - ], - "sampleResponses": [ # Optional. Sample response for the Agent. - { # Sample response that the agent can use. This could be based on applicable instructions and ingested data from other systems. - "responseText": "A String", # Optional. Sample response for Agent in text. - }, - ], - "suggestionEval": { # Self evaluations of the suggestion. # Self evaluation of the suggestion. - "actionActionSuggestionEval": "A String", # Optional. Eval for Agent action suggestion. - "sampleResponseEval": "A String", # Optional. Eval for sample response. - }, - "suggestionReasoning": { # Reasoning for the suggestion. # Reasoning for the suggestion. - "agentActionTaken": "A String", # Optional. The actions that the agent has taken already. - "issueSummary": "A String", # Optional. Summary of the issue. - }, - }, - "freeFormSuggestion": { # Suggestion generated using free form generator. # Optional. Free form suggestion. - "labels": [ # Optional. Labels for the generator. - "A String", - ], - "response": "A String", # Required. Free form suggestion. - }, - "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. - "summarySections": [ # Required. All the parts of generated summary. - { # A component of the generated summary. - "section": "A String", # Required. Name of the section. - "summary": "A String", # Required. Summary text for the section. - }, - ], - }, - }, - }, - "knowledgeAssistResult": { # Response for Knowledge Assist. Contains suggested query and optionally includes an answer for the query. # The Knowledge Assist result. - "suggestedQuery": { # Represents a suggested query. # The query suggested based on the context. Suggestion is made only if it is different from the previous suggestion. - "queryText": "A String", # Suggested query text. - "score": 3.14, # Suggested query score. - }, - "suggestedQueryAnswer": { # Represents an answer from Knowledge. Cuurently supports FAQ and Generative answers. # The answer generated for the suggested query. Whether or not an answer is generated depends on how confident we are about the generated query. - "answerText": "A String", # The piece of text from the `source` that answers this suggested query. - "faqSource": { # Details about source of FAQ answer. # Populated if the prediction came from FAQ. - "document": "A String", # Indicates which Knowledge Document this answer was extracted from. Format: `projects//knowledgeBases//documents/`. - "question": "A String", # The corresponding FAQ question. - }, - "generativeSource": { # Details about source of Generative answer. # Populated if the prediction was Generative. - "snippets": [ # All snippets used for this Generative Prediction, with their source URI and data. - { # Snippet Source for a Generative Prediction. - "document": "A String", # Indicates which Knowledge Document this snippet was extracted from. Format: `projects//knowledgeBases//documents/`. - "text": "A String", # text taken from that URI. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - ], - }, - "intentMatchingSource": { # Details about source of Intent Matching answer. # Populated if the prediction was from intent matching. - "title": "A String", # Title of the document. - "uri": "A String", # URI the data is sourced from. - }, - "matchConfidence": 3.14, # The system's confidence score that this answer is a good match for this conversational query. The range is from 0.0 (completely uncertain) to 1.0 (completely certain). - }, - }, - "knowledgeSearchResult": { # Represents a SearchKnowledge answer. # The Knowledge Search result. - "answer": "A String", # The piece of text from the knowledge base documents that answers the search query - "answerRecord": "A String", # The name of the answer record. Format: `projects//locations//answer Records/` - "answerSources": [ # All sources used to generate the answer. - { # The sources of the answers. - "document": "A String", # The document from which the snippet was extracted. Format: `projects//knowledgeBases//documents/` - "snippet": "A String", # The relevant snippet of the article. - "title": "A String", # The title of the article. - "uri": "A String", # The URI of the article. - }, - ], - "answerType": "A String", # The type of the answer. - "confidenceScore": 3.14, # The confidence score in [0.0, 1.0] range. - }, "smartComposeSuggestion": { # Agent Assist Smart Compose suggestion data. # Agent Assist Smart Compose suggestion data. "confidenceScore": 3.14, # The system's confidence score that this suggestion is a good match for this conversation, ranging from 0.0 (completely uncertain) to 1.0 (completely certain). "metadata": { # Map that contains metadata about the Smart Compose suggestion and the document from which it originates. @@ -3106,6 +2486,10 @@Method Details
"transcriptIndex": 42, # The index in the sequence of transcribed pieces of the conversation where the boundary is located. This index starts at zero. "wordIndex": 42, # The word index of this boundary with respect to the first word in the transcript piece. This index starts at zero. }, + "userInput": { # Explicit input used for generating the answer # Explicit input used for generating the answer + "generatorName": "A String", # The resource name of associated generator. Format: `projects//locations//generators/` + "query": "A String", # Query text. Article Search uses this to store the input query used to generate the search results. + }, }, ], "startTime": "A String", # The time at which the conversation started. diff --git a/docs/dyn/datamigration_v1.projects.locations.migrationJobs.html b/docs/dyn/datamigration_v1.projects.locations.migrationJobs.html index 8dddf96a5fd..a0b02dcdae9 100644 --- a/docs/dyn/datamigration_v1.projects.locations.migrationJobs.html +++ b/docs/dyn/datamigration_v1.projects.locations.migrationJobs.html @@ -212,6 +212,7 @@Method Details
}, }, ], + "useDiffBackup": True or False, # Optional. Enable differential backups. }, "state": "A String", # The current migration job state. "staticIpConnectivity": { # The source database will allow incoming connections from the public IP of the destination database. You can retrieve the public IP of the Cloud SQL instance from the Cloud SQL console or using Cloud SQL APIs. No additional configuration is required. # static ip connectivity data (default, no additional details needed). @@ -476,6 +477,7 @@Method Details
}, }, ], + "useDiffBackup": True or False, # Optional. Enable differential backups. }, "state": "A String", # The current migration job state. "staticIpConnectivity": { # The source database will allow incoming connections from the public IP of the destination database. You can retrieve the public IP of the Cloud SQL instance from the Cloud SQL console or using Cloud SQL APIs. No additional configuration is required. # static ip connectivity data (default, no additional details needed). @@ -622,6 +624,7 @@Method Details
}, }, ], + "useDiffBackup": True or False, # Optional. Enable differential backups. }, "state": "A String", # The current migration job state. "staticIpConnectivity": { # The source database will allow incoming connections from the public IP of the destination database. You can retrieve the public IP of the Cloud SQL instance from the Cloud SQL console or using Cloud SQL APIs. No additional configuration is required. # static ip connectivity data (default, no additional details needed). @@ -729,6 +732,7 @@Method Details
}, }, ], + "useDiffBackup": True or False, # Optional. Enable differential backups. }, "state": "A String", # The current migration job state. "staticIpConnectivity": { # The source database will allow incoming connections from the public IP of the destination database. You can retrieve the public IP of the Cloud SQL instance from the Cloud SQL console or using Cloud SQL APIs. No additional configuration is required. # static ip connectivity data (default, no additional details needed). @@ -1170,6 +1174,7 @@Method Details
}, }, ], + "useDiffBackup": True or False, # Optional. Enable differential backups. }, "state": "A String", # The current migration job state. "staticIpConnectivity": { # The source database will allow incoming connections from the public IP of the destination database. You can retrieve the public IP of the Cloud SQL instance from the Cloud SQL console or using Cloud SQL APIs. No additional configuration is required. # static ip connectivity data (default, no additional details needed). diff --git a/docs/dyn/dataplex_v1.projects.locations.dataScans.html b/docs/dyn/dataplex_v1.projects.locations.dataScans.html index 2b31d365b90..7d51b67f2e3 100644 --- a/docs/dyn/dataplex_v1.projects.locations.dataScans.html +++ b/docs/dyn/dataplex_v1.projects.locations.dataScans.html @@ -240,7 +240,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -272,7 +272,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -347,7 +347,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -506,7 +506,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -656,7 +656,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -688,7 +688,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -763,7 +763,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -988,7 +988,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -1020,7 +1020,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -1095,7 +1095,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -1281,7 +1281,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -1313,7 +1313,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -1388,7 +1388,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -1591,7 +1591,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -1623,7 +1623,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -1698,7 +1698,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. diff --git a/docs/dyn/dataplex_v1.projects.locations.dataScans.jobs.html b/docs/dyn/dataplex_v1.projects.locations.dataScans.jobs.html index df3fe99afd2..e22431801bc 100644 --- a/docs/dyn/dataplex_v1.projects.locations.dataScans.jobs.html +++ b/docs/dyn/dataplex_v1.projects.locations.dataScans.jobs.html @@ -142,7 +142,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -287,7 +287,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -319,7 +319,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -394,7 +394,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -548,7 +548,7 @@Method Details
"rowCount": "A String", # The count of rows processed. "rules": [ # A list of all the rules in a job, and their results. { # DataQualityRuleResult provides a more detailed, per-rule view of the results. - "assertionRowCount": "A String", # Output only. The number of rows returned by the sql statement in the SqlAssertion rule.This field is only valid for SqlAssertion rules. + "assertionRowCount": "A String", # Output only. The number of rows returned by the SQL statement in a SQL assertion rule.This field is only valid for SQL assertion rules. "evaluatedCount": "A String", # The number of rows a rule was evaluated against.This field is only valid for row-level type rules.Evaluated count can be configured to either include all rows (default) - with null rows automatically failing rule evaluation, or exclude null rows from the evaluated_count, by setting ignore_nulls = true. "failingRowsQuery": "A String", # The query to find rows that did not pass this rule.This field is only valid for row-level type rules. "nullCount": "A String", # The number of rows with null values in the specified column. @@ -580,7 +580,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. @@ -655,7 +655,7 @@Method Details
"A String", ], }, - "sqlAssertion": { # Queries for rows returned by the provided SQL statement. If any rows are are returned, this rule fails.The SQL statement needs to use BigQuery standard SQL syntax, and must not contain any semicolons.${data()} can be used to reference the rows being evaluated, i.e. the table after all additional filters (row filters, incremental data filters, sampling) are applied.Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. + "sqlAssertion": { # A SQL statement that is evaluated to return rows that match an invalid state. If any rows are are returned, this rule fails.The SQL statement must use BigQuery standard SQL syntax, and must not contain any semicolons.You can use the data reference parameter ${data()} to reference the source table with all of its precondition filters applied. Examples of precondition filters include row filters, incremental data filters, and sampling. For more information, see Data reference parameter (https://cloud.google.com/dataplex/docs/auto-data-quality-overview#data-reference-parameter).Example: SELECT * FROM ${data()} WHERE price < 0 # Aggregate rule which evaluates the number of rows returned for the provided statement. If any rows are returned, this rule fails. "sqlStatement": "A String", # Optional. The SQL statement. }, "statisticRangeExpectation": { # Evaluates whether the column aggregate statistic lies between a specified range. # Aggregate rule which evaluates whether the column aggregate statistic lies between a specified range. diff --git a/docs/dyn/dataplex_v1.projects.locations.html b/docs/dyn/dataplex_v1.projects.locations.html index 6b9462d01ef..bd73d38aa5b 100644 --- a/docs/dyn/dataplex_v1.projects.locations.html +++ b/docs/dyn/dataplex_v1.projects.locations.html @@ -312,7 +312,7 @@Method Details
"nextPageToken": "A String", # Pagination token. "results": [ # The results matching the search query. { # A single result of a SearchEntries request. - "dataplexEntry": { # An entry is a representation of a data asset which can be described by various metadata. # Entry format of the result. + "dataplexEntry": { # An entry is a representation of a data asset which can be described by various metadata. "aspects": { # Optional. The Aspects attached to the Entry. The format for the key can be one of the following: 1. {projectId}.{locationId}.{aspectTypeId} (if the aspect is attached directly to the entry) 2. {projectId}.{locationId}.{aspectTypeId}@{path} (if the aspect is attached to an entry's path) "a_key": { # An aspect is a single piece of metadata describing an entry. "aspectSource": { # AspectSource contains source system related information for the aspect. diff --git a/docs/dyn/datastream_v1.projects.locations.streams.html b/docs/dyn/datastream_v1.projects.locations.streams.html index 7353070265f..599524fc3b2 100644 --- a/docs/dyn/datastream_v1.projects.locations.streams.html +++ b/docs/dyn/datastream_v1.projects.locations.streams.html @@ -439,6 +439,8 @@Method Details
}, "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` "sqlServerSourceConfig": { # SQLServer data source configuration # SQLServer data source configuration. + "changeTables": { # Configuration to use Change Tables CDC read method. # CDC reader reads from change tables. + }, "excludeObjects": { # SQLServer database structure. # SQLServer objects to exclude from the stream. "schemas": [ # SQLServer schemas in the database server. { # SQLServer schema. @@ -489,6 +491,8 @@Method Details
}, "maxConcurrentBackfillTasks": 42, # Max concurrent backfill tasks. "maxConcurrentCdcTasks": 42, # Max concurrent CDC tasks. + "transactionLogs": { # Configuration to use Transaction Logs CDC read method. # CDC reader reads from transaction logs. + }, }, }, "state": "A String", # The state of the stream. @@ -899,6 +903,8 @@Method Details
}, "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` "sqlServerSourceConfig": { # SQLServer data source configuration # SQLServer data source configuration. + "changeTables": { # Configuration to use Change Tables CDC read method. # CDC reader reads from change tables. + }, "excludeObjects": { # SQLServer database structure. # SQLServer objects to exclude from the stream. "schemas": [ # SQLServer schemas in the database server. { # SQLServer schema. @@ -949,6 +955,8 @@Method Details
}, "maxConcurrentBackfillTasks": 42, # Max concurrent backfill tasks. "maxConcurrentCdcTasks": 42, # Max concurrent CDC tasks. + "transactionLogs": { # Configuration to use Transaction Logs CDC read method. # CDC reader reads from transaction logs. + }, }, }, "state": "A String", # The state of the stream. @@ -1298,6 +1306,8 @@Method Details
}, "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` "sqlServerSourceConfig": { # SQLServer data source configuration # SQLServer data source configuration. + "changeTables": { # Configuration to use Change Tables CDC read method. # CDC reader reads from change tables. + }, "excludeObjects": { # SQLServer database structure. # SQLServer objects to exclude from the stream. "schemas": [ # SQLServer schemas in the database server. { # SQLServer schema. @@ -1348,6 +1358,8 @@Method Details
}, "maxConcurrentBackfillTasks": 42, # Max concurrent backfill tasks. "maxConcurrentCdcTasks": 42, # Max concurrent CDC tasks. + "transactionLogs": { # Configuration to use Transaction Logs CDC read method. # CDC reader reads from transaction logs. + }, }, }, "state": "A String", # The state of the stream. @@ -1704,6 +1716,8 @@Method Details
}, "sourceConnectionProfile": "A String", # Required. Source connection profile resoource. Format: `projects/{project}/locations/{location}/connectionProfiles/{name}` "sqlServerSourceConfig": { # SQLServer data source configuration # SQLServer data source configuration. + "changeTables": { # Configuration to use Change Tables CDC read method. # CDC reader reads from change tables. + }, "excludeObjects": { # SQLServer database structure. # SQLServer objects to exclude from the stream. "schemas": [ # SQLServer schemas in the database server. { # SQLServer schema. @@ -1754,6 +1768,8 @@Method Details
}, "maxConcurrentBackfillTasks": 42, # Max concurrent backfill tasks. "maxConcurrentCdcTasks": 42, # Max concurrent CDC tasks. + "transactionLogs": { # Configuration to use Transaction Logs CDC read method. # CDC reader reads from transaction logs. + }, }, }, "state": "A String", # The state of the stream. diff --git a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html index f072a9b9667..c23e7ee4e92 100644 --- a/docs/dyn/dialogflow_v2.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.conversationProfiles.html @@ -219,6 +219,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -273,6 +276,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -285,7 +291,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -398,6 +404,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -452,6 +461,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -464,7 +476,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -602,6 +614,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -656,6 +671,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -668,7 +686,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -792,6 +810,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -846,6 +867,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -858,7 +882,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -990,6 +1014,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1044,6 +1071,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1056,7 +1086,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -1170,6 +1200,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1224,6 +1257,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1236,7 +1272,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2.projects.generators.html b/docs/dyn/dialogflow_v2.projects.generators.html new file mode 100644 index 00000000000..6b077fe3bab --- /dev/null +++ b/docs/dyn/dialogflow_v2.projects.generators.html @@ -0,0 +1,333 @@ + + + +Dialogflow API . projects . generators
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, generatorId=None, x__xgafv=None)
Creates a generator.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists generators.
+ +Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, generatorId=None, x__xgafv=None)
+Creates a generator. + +Args: + parent: string, Required. The project/location to create generator for. Format: `projects//locations/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + generatorId: string, Optional. The ID to use for the generator, which will become the final component of the generator's resource name. The generator ID must be compliant with the regression fomula `a-zA-Z*` with the characters length in range of [3,64]. If the field is not provided, an Id will be auto-generated. If the field is provided, the caller is resposible for 1. the uniqueness of the ID, otherwise the request will be rejected. 2. the consistency for whether to use custom ID or not under a project to better ensure uniqueness. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists generators. + +Args: + parent: string, Required. The project/location to list generators for. Format: `projects//locations/` (required) + pageSize: integer, Optional. Maximum number of conversation models to return in a single page. Default to 10. + pageToken: string, Optional. The next_page_token value returned from a previous list request. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response of ListGenerators. + "generators": [ # List of generators retrieved. + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + ], + "nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list. +}+++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2.projects.html b/docs/dyn/dialogflow_v2.projects.html index 6cb082c1778..1d689d4ea09 100644 --- a/docs/dyn/dialogflow_v2.projects.html +++ b/docs/dyn/dialogflow_v2.projects.html @@ -104,6 +104,11 @@list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Instance Methods
Returns the conversations Resource.
++
+generators()
+Returns the generators Resource.
+ diff --git a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html index c311eda691d..16ef84b53ad 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2.projects.locations.conversationProfiles.html @@ -219,6 +219,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -273,6 +276,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -285,7 +291,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -398,6 +404,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -452,6 +461,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -464,7 +476,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -602,6 +614,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -656,6 +671,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -668,7 +686,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -792,6 +810,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -846,6 +867,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -858,7 +882,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -990,6 +1014,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1044,6 +1071,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1056,7 +1086,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -1170,6 +1200,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1224,6 +1257,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1236,7 +1272,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2.projects.locations.generators.html b/docs/dyn/dialogflow_v2.projects.locations.generators.html new file mode 100644 index 00000000000..d38271e78e8 --- /dev/null +++ b/docs/dyn/dialogflow_v2.projects.locations.generators.html @@ -0,0 +1,577 @@ + + + +Dialogflow API . projects . locations . generators
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, generatorId=None, x__xgafv=None)
Creates a generator.
+ +Deletes a generator.
+ +Retrieves a generator.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists generators.
+ +Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a generator.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, generatorId=None, x__xgafv=None)
+Creates a generator. + +Args: + parent: string, Required. The project/location to create generator for. Format: `projects//locations/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + generatorId: string, Optional. The ID to use for the generator, which will become the final component of the generator's resource name. The generator ID must be compliant with the regression fomula `a-zA-Z*` with the characters length in range of [3,64]. If the field is not provided, an Id will be auto-generated. If the field is provided, the caller is resposible for 1. the uniqueness of the ID, otherwise the request will be rejected. 2. the consistency for whether to use custom ID or not under a project to better ensure uniqueness. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +delete(name, x__xgafv=None)
+Deletes a generator. + +Args: + name: string, Required. The generator resource name to delete. Format: `projects//locations//generators/` (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } +}+++ +get(name, x__xgafv=None)
+Retrieves a generator. + +Args: + name: string, Required. The generator resource name to retrieve. Format: `projects//locations/`/generators/` (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists generators. + +Args: + parent: string, Required. The project/location to list generators for. Format: `projects//locations/` (required) + pageSize: integer, Optional. Maximum number of conversation models to return in a single page. Default to 10. + pageToken: string, Optional. The next_page_token value returned from a previous list request. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response of ListGenerators. + "generators": [ # List of generators retrieved. + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + ], + "nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list. +}+++ +list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2.projects.locations.html b/docs/dyn/dialogflow_v2.projects.locations.html index 98b78dd3c20..11b9e7ab32f 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.html +++ b/docs/dyn/dialogflow_v2.projects.locations.html @@ -104,6 +104,11 @@patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates a generator. + +Args: + name: string, Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + updateMask: string, Optional. The list of fields to update. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+Instance Methods
Returns the conversations Resource.
++
+generators()
+Returns the generators Resource.
+ @@ -114,6 +119,11 @@Instance Methods
Returns the operations Resource.
+ +Returns the statelessSuggestion Resource.
+ diff --git a/docs/dyn/dialogflow_v2.projects.locations.statelessSuggestion.html b/docs/dyn/dialogflow_v2.projects.locations.statelessSuggestion.html new file mode 100644 index 00000000000..53066f0dc86 --- /dev/null +++ b/docs/dyn/dialogflow_v2.projects.locations.statelessSuggestion.html @@ -0,0 +1,197 @@ + + + +Dialogflow API . projects . locations . statelessSuggestion
+Instance Methods
++
+close()
Close httplib2 connections.
++
+generate(parent, body=None, x__xgafv=None)
Generates and returns a suggestion for a conversation that does not have a resource created for it.
+Method Details
+++ +close()
+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html index 9ec0c2aa413..5f04593c75d 100644 --- a/docs/dyn/dialogflow_v2.projects.locations.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.locations.suggestions.html @@ -159,6 +159,9 @@generate(parent, body=None, x__xgafv=None)
+Generates and returns a suggestion for a conversation that does not have a resource created for it. + +Args: + parent: string, Required. The parent resource to charge for the Suggestion's generation. Format: `projects//locations/`. (required) + body: object, The request body. + The object takes the form of: + +{ # The request message for Conversations.GenerateStatelessSuggestion. + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Context of the conversation, including transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "generator": { # LLM generator. # Uncreated generator. It should be a complete generator that includes all information about the generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + "generatorName": "A String", # The resource name of the existing created generator. Format: `projects//locations//generators/` + "triggerEvents": [ # Optional. A list of trigger events. Generator will be triggered only if it's trigger event is included here. + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response message for Conversations.GenerateStatelessSuggestion. + "generatorSuggestion": { # Suggestion generated using a Generator. # Required. Generated suggestion for a conversation. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, +}+Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -213,6 +216,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -225,7 +231,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2.projects.suggestions.html b/docs/dyn/dialogflow_v2.projects.suggestions.html index 6a8fcbbd97e..786ad5780c6 100644 --- a/docs/dyn/dialogflow_v2.projects.suggestions.html +++ b/docs/dyn/dialogflow_v2.projects.suggestions.html @@ -159,6 +159,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -213,6 +216,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -225,7 +231,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html index 00f23541dab..3630c61330e 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.environments.users.sessions.html @@ -152,6 +152,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html index fc78f69088d..b4eec144ec8 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.agent.sessions.html @@ -152,6 +152,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html index cdd33a755da..a0b7c0002c3 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversationProfiles.html @@ -219,6 +219,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -273,6 +276,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -285,7 +291,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -398,6 +404,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -452,6 +461,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -464,7 +476,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -602,6 +614,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -656,6 +671,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -668,7 +686,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -792,6 +810,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -846,6 +867,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -858,7 +882,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -990,6 +1014,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1044,6 +1071,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1056,7 +1086,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -1170,6 +1200,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1224,6 +1257,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1236,7 +1272,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html index 09b24e78ae6..b5f0c0d787a 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.conversations.participants.html @@ -124,6 +124,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.generators.html b/docs/dyn/dialogflow_v2beta1.projects.generators.html new file mode 100644 index 00000000000..a8e02a245a3 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.generators.html @@ -0,0 +1,333 @@ + + + +Dialogflow API . projects . generators
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, generatorId=None, x__xgafv=None)
Creates a generator.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists generators.
+ +Retrieves the next page of results.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, generatorId=None, x__xgafv=None)
+Creates a generator. + +Args: + parent: string, Required. The project/location to create generator for. Format: `projects//locations/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + generatorId: string, Optional. The ID to use for the generator, which will become the final component of the generator's resource name. The generator ID must be compliant with the regression fomula `a-zA-Z*` with the characters length in range of [3,64]. If the field is not provided, an Id will be auto-generated. If the field is provided, the caller is resposible for 1. the uniqueness of the ID, otherwise the request will be rejected. 2. the consistency for whether to use custom ID or not under a project to better ensure uniqueness. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists generators. + +Args: + parent: string, Required. The project/location to list generators for. Format: `projects//locations/` (required) + pageSize: integer, Optional. Maximum number of conversation models to return in a single page. Default to 10. + pageToken: string, Optional. The next_page_token value returned from a previous list request. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response of ListGenerators. + "generators": [ # List of generators retrieved. + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + ], + "nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list. +}+++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2beta1.projects.html b/docs/dyn/dialogflow_v2beta1.projects.html index 13f3be6396e..5508dae56c2 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.html +++ b/docs/dyn/dialogflow_v2beta1.projects.html @@ -94,6 +94,11 @@list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++Instance Methods
Returns the conversations Resource.
++
+generators()
+Returns the generators Resource.
+ diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html index d8c40ef4e85..96597bba693 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.environments.users.sessions.html @@ -152,6 +152,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html index 2063ce7e05b..f07baee3476 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.agent.sessions.html @@ -152,6 +152,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html index a6d871e7ed5..d8284cfe6b5 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversationProfiles.html @@ -219,6 +219,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -273,6 +276,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -285,7 +291,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -398,6 +404,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -452,6 +461,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -464,7 +476,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -602,6 +614,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -656,6 +671,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -668,7 +686,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -792,6 +810,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -846,6 +867,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -858,7 +882,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -990,6 +1014,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1044,6 +1071,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1056,7 +1086,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. @@ -1170,6 +1200,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -1224,6 +1257,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -1236,7 +1272,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html index ccdccc05718..f1f5c2c0b1b 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.conversations.participants.html @@ -124,6 +124,7 @@Method Details
"noBargeInDuration": "A String", # Duration that is not eligible for barge-in at the beginning of the input audio. "totalDuration": "A String", # Total duration for the playback at the beginning of the input audio. }, + "defaultNoSpeechTimeout": "A String", # If set, use this no-speech timeout when the agent does not provide a no-speech timeout itself. "disableNoSpeechRecognizedEvent": True or False, # Only used in Participants.AnalyzeContent and Participants.StreamingAnalyzeContent. If `false` and recognition doesn't return any result, trigger `NO_SPEECH_RECOGNIZED` event to Dialogflow agent. "enableAutomaticPunctuation": True or False, # Enable automatic punctuation option at the speech backend. "enableWordInfo": True or False, # If `true`, Dialogflow returns SpeechWordInfo in StreamingRecognitionResult with information about the recognized speech words, e.g. start and end time offsets. If false or unspecified, Speech doesn't return any word-level information. diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.generators.html b/docs/dyn/dialogflow_v2beta1.projects.locations.generators.html new file mode 100644 index 00000000000..d94baf7cfc6 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.generators.html @@ -0,0 +1,577 @@ + + + +Dialogflow API . projects . locations . generators
+Instance Methods
++
+close()
Close httplib2 connections.
++
+create(parent, body=None, generatorId=None, x__xgafv=None)
Creates a generator.
+ +Deletes a generator.
+ +Retrieves a generator.
++
+list(parent, pageSize=None, pageToken=None, x__xgafv=None)
Lists generators.
+ +Retrieves the next page of results.
++
+patch(name, body=None, updateMask=None, x__xgafv=None)
Updates a generator.
+Method Details
+++ +close()
+Close httplib2 connections.+++ +create(parent, body=None, generatorId=None, x__xgafv=None)
+Creates a generator. + +Args: + parent: string, Required. The project/location to create generator for. Format: `projects//locations/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + generatorId: string, Optional. The ID to use for the generator, which will become the final component of the generator's resource name. The generator ID must be compliant with the regression fomula `a-zA-Z*` with the characters length in range of [3,64]. If the field is not provided, an Id will be auto-generated. If the field is provided, the caller is resposible for 1. the uniqueness of the ID, otherwise the request will be rejected. 2. the consistency for whether to use custom ID or not under a project to better ensure uniqueness. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +delete(name, x__xgafv=None)
+Deletes a generator. + +Args: + name: string, Required. The generator resource name to delete. Format: `projects//locations//generators/` (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } +}+++ +get(name, x__xgafv=None)
+Retrieves a generator. + +Args: + name: string, Required. The generator resource name to retrieve. Format: `projects//locations/`/generators/` (required) + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+++ +list(parent, pageSize=None, pageToken=None, x__xgafv=None)
+Lists generators. + +Args: + parent: string, Required. The project/location to list generators for. Format: `projects//locations/` (required) + pageSize: integer, Optional. Maximum number of conversation models to return in a single page. Default to 10. + pageToken: string, Optional. The next_page_token value returned from a previous list request. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # Response of ListGenerators. + "generators": [ # List of generators retrieved. + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + ], + "nextPageToken": "A String", # Token to retrieve the next page of results, or empty if there are no more results in the list. +}+++ +list_next()
+Retrieves the next page of results. + + Args: + previous_request: The request for the previous page. (required) + previous_response: The response from the request for the previous page. (required) + + Returns: + A request object that you can call 'execute()' on to request the next + page. Returns None if there are no more items in the collection. ++++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.html b/docs/dyn/dialogflow_v2beta1.projects.locations.html index 94406a0d441..ea91938d5ad 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.html @@ -94,6 +94,11 @@patch(name, body=None, updateMask=None, x__xgafv=None)
+Updates a generator. + +Args: + name: string, Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` (required) + body: object, The request body. + The object takes the form of: + +{ # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +} + + updateMask: string, Optional. The list of fields to update. + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # LLM generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. +}+Instance Methods
Returns the conversations Resource.
++
+generators()
+Returns the generators Resource.
+ @@ -104,6 +109,11 @@Instance Methods
Returns the operations Resource.
+ +Returns the statelessSuggestion Resource.
+ diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.statelessSuggestion.html b/docs/dyn/dialogflow_v2beta1.projects.locations.statelessSuggestion.html new file mode 100644 index 00000000000..5366a3c5c51 --- /dev/null +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.statelessSuggestion.html @@ -0,0 +1,197 @@ + + + +Dialogflow API . projects . locations . statelessSuggestion
+Instance Methods
++
+close()
Close httplib2 connections.
++
+generate(parent, body=None, x__xgafv=None)
Generates and returns a suggestion for a conversation that does not have a resource created for it.
+Method Details
+++ +close()
+Close httplib2 connections.+++ + \ No newline at end of file diff --git a/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html index 233eb607534..462615c5026 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.locations.suggestions.html @@ -159,6 +159,9 @@generate(parent, body=None, x__xgafv=None)
+Generates and returns a suggestion for a conversation that does not have a resource created for it. + +Args: + parent: string, Required. The parent resource to charge for the Suggestion's generation. Format: `projects//locations/`. (required) + body: object, The request body. + The object takes the form of: + +{ # The request message for Conversations.GenerateStatelessSuggestion. + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Context of the conversation, including transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "generator": { # LLM generator. # Uncreated generator. It should be a complete generator that includes all information about the generator. + "createTime": "A String", # Output only. Creation time of this generator. + "description": "A String", # Optional. Human readable description of the generator. + "inferenceParameter": { # The parameters of inference. # Optional. Inference parameters for this generator. + "maxOutputTokens": 42, # Optional. Maximum number of the output tokens for the generator. + "temperature": 3.14, # Optional. Controls the randomness of LLM predictions. Low temperature = less random. High temperature = more random. If unset (or 0), uses a default value of 0. + "topK": 42, # Optional. Top-k changes how the model selects tokens for output. A top-k of 1 means the selected token is the most probable among all tokens in the model's vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature). For each token selection step, the top K tokens with the highest probabilities are sampled. Then tokens are further filtered based on topP with the final token selected using temperature sampling. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [1, 40], default to 40. + "topP": 3.14, # Optional. Top-p changes how the model selects tokens for output. Tokens are selected from most K (see topK parameter) probable to least until the sum of their probabilities equals the top-p value. For example, if tokens A, B, and C have a probability of 0.3, 0.2, and 0.1 and the top-p value is 0.5, then the model will select either A or B as the next token (using temperature) and doesn't consider C. The default top-p value is 0.95. Specify a lower value for less random responses and a higher value for more random responses. Acceptable value is [0.0, 1.0], default to 0.95. + }, + "name": "A String", # Output only. Identifier. The resource name of the generator. Format: `projects//locations//generators/` + "summarizationContext": { # Summarization context that customer can configure. # Input of prebuilt Summarization feature. + "fewShotExamples": [ # Optional. List of few shot examples. + { # Providing examples in the generator (i.e. building a few-shot generator) helps convey the desired format of the LLM response. NEXT_ID: 10 + "conversationContext": { # Context of the conversation, including transcripts. # Optional. Conversation transcripts. + "messageEntries": [ # Optional. List of message transcripts in the conversation. + { # Represents a message entry of a conversation. + "createTime": "A String", # Optional. Create time of the message entry. + "languageCode": "A String", # Optional. The language of the text. See [Language Support](https://cloud.google.com/dialogflow/docs/reference/language) for a list of the currently supported language codes. + "role": "A String", # Optional. Participant role of the message. + "text": "A String", # Optional. Transcript content of the message. + }, + ], + }, + "extraInfo": { # Optional. Key is the placeholder field name in input, value is the value of the placeholder. E.g. instruction contains "@price", and ingested data has <"price", "10"> + "a_key": "A String", + }, + "output": { # Suggestion generated using a Generator. # Required. Example output of the model. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, + "summarizationSectionList": { # List of summarization sections. # Summarization sections. + "summarizationSections": [ # Optional. Summarization sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + }, + }, + ], + "outputLanguageCode": "A String", # Optional. The target language of the generated summary. The language code for conversation will be used if this field is empty. Supported 2.0 and later versions. + "summarizationSections": [ # Optional. List of sections. Note it contains both predefined section sand customer defined sections. + { # Represents the section of summarization. + "definition": "A String", # Optional. Definition of the section, for example, "what the customer needs help with or has question about." + "key": "A String", # Optional. Name of the section, for example, "situation". + "type": "A String", # Optional. Type of the summarization section. + }, + ], + "version": "A String", # Optional. Version of the feature. If not set, default to latest version. Current candidates are ["1.0"]. + }, + "triggerEvent": "A String", # Optional. The trigger event of the generator. It defines when the generator is triggered in a conversation. + "updateTime": "A String", # Output only. Update time of this generator. + }, + "generatorName": "A String", # The resource name of the existing created generator. Format: `projects//locations//generators/` + "triggerEvents": [ # Optional. A list of trigger events. Generator will be triggered only if it's trigger event is included here. + "A String", + ], +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # The response message for Conversations.GenerateStatelessSuggestion. + "generatorSuggestion": { # Suggestion generated using a Generator. # Required. Generated suggestion for a conversation. + "summarySuggestion": { # Suggested summary of the conversation. # Optional. Suggested summary. + "summarySections": [ # Required. All the parts of generated summary. + { # A component of the generated summary. + "section": "A String", # Required. Name of the section. + "summary": "A String", # Required. Summary text for the section. + }, + ], + }, + }, +}+Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -213,6 +216,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -225,7 +231,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/dialogflow_v2beta1.projects.suggestions.html b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html index fa17eb6b68c..a06739ce92b 100644 --- a/docs/dyn/dialogflow_v2beta1.projects.suggestions.html +++ b/docs/dyn/dialogflow_v2beta1.projects.suggestions.html @@ -159,6 +159,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "humanAgentSuggestionConfig": { # Detail human agent assistant config. # Configuration for agent assistance of human agent participant. @@ -213,6 +216,9 @@Method Details
}, }, ], + "generators": [ # Optional. List of various generator resource names used in the conversation profile. + "A String", + ], "groupSuggestionResponses": True or False, # If `group_suggestion_responses` is false, and there are multiple `feature_configs` in `event based suggestion` or StreamingAnalyzeContent, we will try to deliver suggestions to customers as soon as we get new suggestion. Different type of suggestions based on the same context will be in separate Pub/Sub event or `StreamingAnalyzeContentResponse`. If `group_suggestion_responses` set to true. All the suggestions to the same participant based on the same context will be grouped into a single Pub/Sub event or StreamingAnalyzeContentResponse. }, "messageAnalysisConfig": { # Configuration for analyses to run on each conversation message. # Configuration for message analysis. @@ -225,7 +231,7 @@Method Details
}, }, "humanAgentHandoffConfig": { # Defines the hand off to a live agent, typically on which external agent service provider to connect to a conversation. Currently, this feature is not general available, please contact Google to get access. # Configuration for connecting to a live agent. Currently, this feature is not general available, please contact Google to get access. - "livePersonConfig": { # Configuration specific to LivePerson (https://www.liveperson.com). # Uses LivePerson (https://www.liveperson.com). + "livePersonConfig": { # Configuration specific to [LivePerson](https://www.liveperson.com). # Uses [LivePerson](https://www.liveperson.com). "accountNumber": "A String", # Required. Account number of the LivePerson account to connect. This is the account number you input at the login page. }, "salesforceLiveAgentConfig": { # Configuration specific to Salesforce Live Agent. # Uses Salesforce Live Agent. diff --git a/docs/dyn/discoveryengine_v1.projects.html b/docs/dyn/discoveryengine_v1.projects.html index e66ad345760..a0cab0f983d 100644 --- a/docs/dyn/discoveryengine_v1.projects.html +++ b/docs/dyn/discoveryengine_v1.projects.html @@ -87,10 +87,56 @@Instance Methods
Close httplib2 connections.
++
+provision(name, body=None, x__xgafv=None)
Provisions the project resource. During the process, related systems will get prepared and initialized. Caller must read the [Terms for data use](https://cloud.google.com/retail/data-use-terms), and optionally specify in request to provide consent to that service terms.
Method Details
+close()
Close httplib2 connections.++provision(name, body=None, x__xgafv=None)
+Provisions the project resource. During the process, related systems will get prepared and initialized. Caller must read the [Terms for data use](https://cloud.google.com/retail/data-use-terms), and optionally specify in request to provide consent to that service terms. + +Args: + name: string, Required. Full resource name of a Project, such as `projects/{project_id_or_number}`. (required) + body: object, The request body. + The object takes the form of: + +{ # Request for ProjectService.ProvisionProject method. + "acceptDataUseTerms": True or False, # Required. Set to `true` to specify that caller has read and would like to give consent to the [Terms for data use](https://cloud.google.com/retail/data-use-terms). + "dataUseTermsVersion": "A String", # Required. The version of the [Terms for data use](https://cloud.google.com/retail/data-use-terms) that caller has read and would like to give consent to. Acceptable version is `2022-11-23`, and this may change over time. +} + + x__xgafv: string, V1 error format. + Allowed values + 1 - v1 error format + 2 - v2 error format + +Returns: + An object of the form: + + { # This resource represents a long-running operation that is the result of a network API call. + "done": True or False, # If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available. + "error": { # The `Status` type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by [gRPC](https://github.com/grpc). Each `Status` message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the [API Design Guide](https://cloud.google.com/apis/design/errors). # The error result of the operation in case of failure or cancellation. + "code": 42, # The status code, which should be an enum value of google.rpc.Code. + "details": [ # A list of messages that carry the error details. There is a common set of message types for APIs to use. + { + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + ], + "message": "A String", # A developer-facing error message, which should be in English. Any user-facing error message should be localized and sent in the google.rpc.Status.details field, or localized by the client. + }, + "metadata": { # Service-specific metadata associated with the operation. It typically contains progress information and common metadata such as create time. Some services might not provide such metadata. Any method that returns a long-running operation should document the metadata type, if any. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, + "name": "A String", # The server-assigned name, which is only unique within the same service that originally returns it. If you use the default HTTP mapping, the `name` should be a resource name ending with `operations/{unique_id}`. + "response": { # The normal, successful response of the operation. If the original method returns no data on success, such as `Delete`, the response is `google.protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`, the response should be the resource. For other methods, the response should have the type `XxxResponse`, where `Xxx` is the original method name. For example, if the original method name is `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. + "a_key": "", # Properties of the object. Contains field @type with type URL. + }, +}+